{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 500, "global_step": 6576, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006083650190114068, "grad_norm": 194.0, "learning_rate": 0.0, "loss": 13.3208, "num_input_tokens_seen": 131072, "step": 1, "train_runtime": 41.7221, "train_tokens_per_second": 3141.549 }, { "epoch": 0.006083650190114068, "grad_norm": 15.875, "learning_rate": 2.272727272727273e-05, "loss": 12.4449, "num_input_tokens_seen": 1310720, "step": 10, "train_runtime": 360.3457, "train_tokens_per_second": 3637.396 }, { "epoch": 0.012167300380228136, "grad_norm": 2.359375, "learning_rate": 4.797979797979798e-05, "loss": 11.0058, "num_input_tokens_seen": 2621440, "step": 20, "train_runtime": 844.7738, "train_tokens_per_second": 3103.126 }, { "epoch": 0.018250950570342206, "grad_norm": 4.75, "learning_rate": 7.323232323232324e-05, "loss": 9.7781, "num_input_tokens_seen": 3932160, "step": 30, "train_runtime": 1327.8156, "train_tokens_per_second": 2961.375 }, { "epoch": 0.024334600760456272, "grad_norm": 4.9375, "learning_rate": 9.848484848484848e-05, "loss": 8.6293, "num_input_tokens_seen": 5242880, "step": 40, "train_runtime": 1818.9813, "train_tokens_per_second": 2882.317 }, { "epoch": 0.030418250950570342, "grad_norm": 5.75, "learning_rate": 0.00012373737373737374, "loss": 8.3717, "num_input_tokens_seen": 6553600, "step": 50, "train_runtime": 2311.0689, "train_tokens_per_second": 2835.744 }, { "epoch": 0.03650190114068441, "grad_norm": 4.25, "learning_rate": 0.00014898989898989897, "loss": 8.144, "num_input_tokens_seen": 7864320, "step": 60, "train_runtime": 2799.9586, "train_tokens_per_second": 2808.727 }, { "epoch": 0.04258555133079848, "grad_norm": 5.65625, "learning_rate": 0.00017424242424242425, "loss": 7.9474, "num_input_tokens_seen": 9175040, "step": 70, "train_runtime": 3293.677, "train_tokens_per_second": 2785.653 }, { "epoch": 0.048669201520912544, "grad_norm": 6.28125, "learning_rate": 0.0001994949494949495, "loss": 7.6349, "num_input_tokens_seen": 10485760, "step": 80, "train_runtime": 3748.5901, "train_tokens_per_second": 2797.254 }, { "epoch": 0.05475285171102662, "grad_norm": 2.140625, "learning_rate": 0.00022474747474747475, "loss": 7.4784, "num_input_tokens_seen": 11796480, "step": 90, "train_runtime": 4186.4419, "train_tokens_per_second": 2817.782 }, { "epoch": 0.060836501901140684, "grad_norm": 2.28125, "learning_rate": 0.00025, "loss": 7.136, "num_input_tokens_seen": 13107200, "step": 100, "train_runtime": 4674.9867, "train_tokens_per_second": 2803.687 }, { "epoch": 0.06692015209125475, "grad_norm": 2.15625, "learning_rate": 0.00027525252525252526, "loss": 6.9612, "num_input_tokens_seen": 14417920, "step": 110, "train_runtime": 5176.2495, "train_tokens_per_second": 2785.399 }, { "epoch": 0.07300380228136882, "grad_norm": 3.015625, "learning_rate": 0.0003005050505050505, "loss": 6.8282, "num_input_tokens_seen": 15728640, "step": 120, "train_runtime": 5683.7739, "train_tokens_per_second": 2767.288 }, { "epoch": 0.07908745247148288, "grad_norm": 1.9140625, "learning_rate": 0.00032575757575757576, "loss": 6.6848, "num_input_tokens_seen": 17039360, "step": 130, "train_runtime": 6191.256, "train_tokens_per_second": 2752.165 }, { "epoch": 0.08517110266159696, "grad_norm": 2.390625, "learning_rate": 0.000351010101010101, "loss": 6.4192, "num_input_tokens_seen": 18350080, "step": 140, "train_runtime": 6698.3992, "train_tokens_per_second": 2739.472 }, { "epoch": 0.09125475285171103, "grad_norm": 1.890625, "learning_rate": 0.00037626262626262627, "loss": 6.1528, "num_input_tokens_seen": 19660800, "step": 150, "train_runtime": 7206.1898, "train_tokens_per_second": 2728.321 }, { "epoch": 0.09733840304182509, "grad_norm": 2.953125, "learning_rate": 0.0004015151515151515, "loss": 6.1066, "num_input_tokens_seen": 20971520, "step": 160, "train_runtime": 7714.79, "train_tokens_per_second": 2718.353 }, { "epoch": 0.10342205323193916, "grad_norm": 2.609375, "learning_rate": 0.00042676767676767677, "loss": 5.8847, "num_input_tokens_seen": 22282240, "step": 170, "train_runtime": 8220.9522, "train_tokens_per_second": 2710.421 }, { "epoch": 0.10950570342205324, "grad_norm": 6.0, "learning_rate": 0.0004520202020202021, "loss": 6.1618, "num_input_tokens_seen": 23592960, "step": 180, "train_runtime": 8731.5822, "train_tokens_per_second": 2702.026 }, { "epoch": 0.1155893536121673, "grad_norm": 7.96875, "learning_rate": 0.0004772727272727273, "loss": 6.0654, "num_input_tokens_seen": 24903680, "step": 190, "train_runtime": 9242.7977, "train_tokens_per_second": 2694.388 }, { "epoch": 0.12167300380228137, "grad_norm": 3.25, "learning_rate": 0.0004999999696722131, "loss": 5.6755, "num_input_tokens_seen": 26214400, "step": 200, "train_runtime": 9756.404, "train_tokens_per_second": 2686.892 }, { "epoch": 0.12775665399239544, "grad_norm": 123.5, "learning_rate": 0.0004999963303466888, "loss": 5.8267, "num_input_tokens_seen": 27525120, "step": 210, "train_runtime": 10265.9757, "train_tokens_per_second": 2681.199 }, { "epoch": 0.1338403041825095, "grad_norm": 4.78125, "learning_rate": 0.0004999866255649598, "loss": 5.7143, "num_input_tokens_seen": 28835840, "step": 220, "train_runtime": 10768.3145, "train_tokens_per_second": 2677.842 }, { "epoch": 0.1399239543726236, "grad_norm": 3.1875, "learning_rate": 0.000499970855562485, "loss": 5.4202, "num_input_tokens_seen": 30146560, "step": 230, "train_runtime": 11276.6, "train_tokens_per_second": 2673.373 }, { "epoch": 0.14600760456273765, "grad_norm": 2.640625, "learning_rate": 0.0004999490207218795, "loss": 5.2555, "num_input_tokens_seen": 31457280, "step": 240, "train_runtime": 11787.6318, "train_tokens_per_second": 2668.668 }, { "epoch": 0.1520912547528517, "grad_norm": 1.875, "learning_rate": 0.0004999211215729038, "loss": 5.1247, "num_input_tokens_seen": 32768000, "step": 250, "train_runtime": 12298.0499, "train_tokens_per_second": 2664.487 }, { "epoch": 0.15817490494296577, "grad_norm": 2.015625, "learning_rate": 0.0004998871587924522, "loss": 5.1095, "num_input_tokens_seen": 34078720, "step": 260, "train_runtime": 12764.4114, "train_tokens_per_second": 2669.823 }, { "epoch": 0.16425855513307985, "grad_norm": 2.53125, "learning_rate": 0.000499847133204536, "loss": 4.9468, "num_input_tokens_seen": 35389440, "step": 270, "train_runtime": 13201.0465, "train_tokens_per_second": 2680.806 }, { "epoch": 0.1703422053231939, "grad_norm": 1.609375, "learning_rate": 0.000499801045780263, "loss": 4.8639, "num_input_tokens_seen": 36700160, "step": 280, "train_runtime": 13643.729, "train_tokens_per_second": 2689.892 }, { "epoch": 0.17642585551330797, "grad_norm": 1.3984375, "learning_rate": 0.0004997488976378147, "loss": 4.785, "num_input_tokens_seen": 38010880, "step": 290, "train_runtime": 14087.3187, "train_tokens_per_second": 2698.234 }, { "epoch": 0.18250950570342206, "grad_norm": 1.9140625, "learning_rate": 0.0004996906900424189, "loss": 4.7422, "num_input_tokens_seen": 39321600, "step": 300, "train_runtime": 14530.9563, "train_tokens_per_second": 2706.057 }, { "epoch": 0.18859315589353612, "grad_norm": 1.484375, "learning_rate": 0.0004996264244063187, "loss": 4.6158, "num_input_tokens_seen": 40632320, "step": 310, "train_runtime": 14975.051, "train_tokens_per_second": 2713.334 }, { "epoch": 0.19467680608365018, "grad_norm": 1.6953125, "learning_rate": 0.0004995561022887386, "loss": 4.5245, "num_input_tokens_seen": 41943040, "step": 320, "train_runtime": 15419.5208, "train_tokens_per_second": 2720.126 }, { "epoch": 0.20076045627376427, "grad_norm": 1.0390625, "learning_rate": 0.0004994797253958465, "loss": 4.4933, "num_input_tokens_seen": 43253760, "step": 330, "train_runtime": 15930.5389, "train_tokens_per_second": 2715.147 }, { "epoch": 0.20684410646387832, "grad_norm": 1.2734375, "learning_rate": 0.0004993972955807125, "loss": 4.4701, "num_input_tokens_seen": 44564480, "step": 340, "train_runtime": 16443.622, "train_tokens_per_second": 2710.138 }, { "epoch": 0.21292775665399238, "grad_norm": 1.7890625, "learning_rate": 0.0004993088148432637, "loss": 4.4604, "num_input_tokens_seen": 45875200, "step": 350, "train_runtime": 16954.6894, "train_tokens_per_second": 2705.753 }, { "epoch": 0.21901140684410647, "grad_norm": 1.3828125, "learning_rate": 0.0004992142853302356, "loss": 4.3468, "num_input_tokens_seen": 47185920, "step": 360, "train_runtime": 17465.395, "train_tokens_per_second": 2701.681 }, { "epoch": 0.22509505703422053, "grad_norm": 2.015625, "learning_rate": 0.0004991137093351205, "loss": 4.3054, "num_input_tokens_seen": 48496640, "step": 370, "train_runtime": 17969.6261, "train_tokens_per_second": 2698.812 }, { "epoch": 0.2311787072243346, "grad_norm": 1.0390625, "learning_rate": 0.0004990070892981112, "loss": 4.2011, "num_input_tokens_seen": 49807360, "step": 380, "train_runtime": 18474.7265, "train_tokens_per_second": 2695.973 }, { "epoch": 0.23726235741444868, "grad_norm": 3.46875, "learning_rate": 0.0004988944278060427, "loss": 4.3475, "num_input_tokens_seen": 51118080, "step": 390, "train_runtime": 18979.6402, "train_tokens_per_second": 2693.311 }, { "epoch": 0.24334600760456274, "grad_norm": 3.65625, "learning_rate": 0.0004987757275923281, "loss": 4.3356, "num_input_tokens_seen": 52428800, "step": 400, "train_runtime": 19484.7862, "train_tokens_per_second": 2690.756 }, { "epoch": 0.2494296577946768, "grad_norm": 10.5625, "learning_rate": 0.0004986509915368937, "loss": 4.186, "num_input_tokens_seen": 53739520, "step": 410, "train_runtime": 19989.039, "train_tokens_per_second": 2688.449 }, { "epoch": 0.2555133079847909, "grad_norm": 51.0, "learning_rate": 0.0004985202226661082, "loss": 4.6079, "num_input_tokens_seen": 55050240, "step": 420, "train_runtime": 20493.286, "train_tokens_per_second": 2686.257 }, { "epoch": 0.26159695817490497, "grad_norm": 9.6875, "learning_rate": 0.0004983834241527096, "loss": 4.3251, "num_input_tokens_seen": 56360960, "step": 430, "train_runtime": 20997.3631, "train_tokens_per_second": 2684.192 }, { "epoch": 0.267680608365019, "grad_norm": 1.859375, "learning_rate": 0.0004982405993157283, "loss": 4.0773, "num_input_tokens_seen": 57671680, "step": 440, "train_runtime": 21503.8583, "train_tokens_per_second": 2681.922 }, { "epoch": 0.2737642585551331, "grad_norm": 1.1328125, "learning_rate": 0.0004980917516204061, "loss": 4.0371, "num_input_tokens_seen": 58982400, "step": 450, "train_runtime": 22012.6337, "train_tokens_per_second": 2679.479 }, { "epoch": 0.2798479087452472, "grad_norm": 1.421875, "learning_rate": 0.0004979368846781129, "loss": 4.065, "num_input_tokens_seen": 60293120, "step": 460, "train_runtime": 22521.0154, "train_tokens_per_second": 2677.194 }, { "epoch": 0.2859315589353612, "grad_norm": 1.6015625, "learning_rate": 0.0004977760022462584, "loss": 4.0267, "num_input_tokens_seen": 61603840, "step": 470, "train_runtime": 23025.7928, "train_tokens_per_second": 2675.428 }, { "epoch": 0.2920152091254753, "grad_norm": 0.94140625, "learning_rate": 0.0004976091082282013, "loss": 3.9963, "num_input_tokens_seen": 62914560, "step": 480, "train_runtime": 23530.8991, "train_tokens_per_second": 2673.7 }, { "epoch": 0.2980988593155893, "grad_norm": 1.1875, "learning_rate": 0.0004974362066731545, "loss": 4.0162, "num_input_tokens_seen": 64225280, "step": 490, "train_runtime": 24031.3065, "train_tokens_per_second": 2672.567 }, { "epoch": 0.3041825095057034, "grad_norm": 1.125, "learning_rate": 0.0004972573017760867, "loss": 3.9844, "num_input_tokens_seen": 65536000, "step": 500, "train_runtime": 24533.3589, "train_tokens_per_second": 2671.302 }, { "epoch": 0.3102661596958175, "grad_norm": 1.40625, "learning_rate": 0.000497072397877621, "loss": 3.999, "num_input_tokens_seen": 66846720, "step": 510, "train_runtime": 25032.6558, "train_tokens_per_second": 2670.381 }, { "epoch": 0.31634980988593153, "grad_norm": 1.109375, "learning_rate": 0.0004968814994639292, "loss": 3.9714, "num_input_tokens_seen": 68157440, "step": 520, "train_runtime": 25526.0535, "train_tokens_per_second": 2670.113 }, { "epoch": 0.3224334600760456, "grad_norm": 0.78515625, "learning_rate": 0.0004966846111666232, "loss": 3.9438, "num_input_tokens_seen": 69468160, "step": 530, "train_runtime": 26024.3771, "train_tokens_per_second": 2669.35 }, { "epoch": 0.3285171102661597, "grad_norm": 0.7109375, "learning_rate": 0.0004964817377626425, "loss": 3.8985, "num_input_tokens_seen": 70778880, "step": 540, "train_runtime": 26527.3154, "train_tokens_per_second": 2668.151 }, { "epoch": 0.33460076045627374, "grad_norm": 0.6875, "learning_rate": 0.0004962728841741383, "loss": 3.926, "num_input_tokens_seen": 72089600, "step": 550, "train_runtime": 27037.0928, "train_tokens_per_second": 2666.322 }, { "epoch": 0.3406844106463878, "grad_norm": 0.859375, "learning_rate": 0.000496058055468354, "loss": 3.8868, "num_input_tokens_seen": 73400320, "step": 560, "train_runtime": 27544.463, "train_tokens_per_second": 2664.794 }, { "epoch": 0.3467680608365019, "grad_norm": 1.046875, "learning_rate": 0.0004958372568575029, "loss": 3.877, "num_input_tokens_seen": 74711040, "step": 570, "train_runtime": 28055.7086, "train_tokens_per_second": 2662.953 }, { "epoch": 0.35285171102661594, "grad_norm": 0.6875, "learning_rate": 0.0004956104936986405, "loss": 3.8675, "num_input_tokens_seen": 76021760, "step": 580, "train_runtime": 28563.1363, "train_tokens_per_second": 2661.534 }, { "epoch": 0.35893536121673003, "grad_norm": 0.77734375, "learning_rate": 0.0004953777714935359, "loss": 3.8062, "num_input_tokens_seen": 77332480, "step": 590, "train_runtime": 29085.1121, "train_tokens_per_second": 2658.834 }, { "epoch": 0.3650190114068441, "grad_norm": 0.63671875, "learning_rate": 0.000495139095888537, "loss": 3.7848, "num_input_tokens_seen": 78643200, "step": 600, "train_runtime": 29603.854, "train_tokens_per_second": 2656.519 }, { "epoch": 0.37110266159695815, "grad_norm": 0.578125, "learning_rate": 0.0004948944726744348, "loss": 3.8038, "num_input_tokens_seen": 79953920, "step": 610, "train_runtime": 30112.8786, "train_tokens_per_second": 2655.14 }, { "epoch": 0.37718631178707224, "grad_norm": 0.87109375, "learning_rate": 0.0004946439077863219, "loss": 3.8141, "num_input_tokens_seen": 81264640, "step": 620, "train_runtime": 30628.1397, "train_tokens_per_second": 2653.267 }, { "epoch": 0.3832699619771863, "grad_norm": 0.59765625, "learning_rate": 0.000494387407303449, "loss": 3.7289, "num_input_tokens_seen": 82575360, "step": 630, "train_runtime": 31145.1722, "train_tokens_per_second": 2651.305 }, { "epoch": 0.38935361216730036, "grad_norm": 0.62109375, "learning_rate": 0.0004941249774490771, "loss": 3.7867, "num_input_tokens_seen": 83886080, "step": 640, "train_runtime": 31656.3857, "train_tokens_per_second": 2649.894 }, { "epoch": 0.39543726235741444, "grad_norm": 0.5703125, "learning_rate": 0.0004938566245903269, "loss": 3.7019, "num_input_tokens_seen": 85196800, "step": 650, "train_runtime": 32167.9519, "train_tokens_per_second": 2648.499 }, { "epoch": 0.40152091254752853, "grad_norm": 0.62109375, "learning_rate": 0.000493582355238024, "loss": 3.7298, "num_input_tokens_seen": 86507520, "step": 660, "train_runtime": 32673.7302, "train_tokens_per_second": 2647.617 }, { "epoch": 0.40760456273764256, "grad_norm": 0.61328125, "learning_rate": 0.0004933021760465411, "loss": 3.6524, "num_input_tokens_seen": 87818240, "step": 670, "train_runtime": 33180.4976, "train_tokens_per_second": 2646.682 }, { "epoch": 0.41368821292775665, "grad_norm": 0.58984375, "learning_rate": 0.0004930160938136364, "loss": 3.6602, "num_input_tokens_seen": 89128960, "step": 680, "train_runtime": 33688.0024, "train_tokens_per_second": 2645.718 }, { "epoch": 0.41977186311787074, "grad_norm": 0.5390625, "learning_rate": 0.0004927241154802888, "loss": 3.6685, "num_input_tokens_seen": 90439680, "step": 690, "train_runtime": 34211.2177, "train_tokens_per_second": 2643.568 }, { "epoch": 0.42585551330798477, "grad_norm": 0.5, "learning_rate": 0.0004924262481305295, "loss": 3.6544, "num_input_tokens_seen": 91750400, "step": 700, "train_runtime": 34733.1748, "train_tokens_per_second": 2641.578 }, { "epoch": 0.43193916349809885, "grad_norm": 0.49609375, "learning_rate": 0.0004921224989912701, "loss": 3.6194, "num_input_tokens_seen": 93061120, "step": 710, "train_runtime": 35254.3191, "train_tokens_per_second": 2639.708 }, { "epoch": 0.43802281368821294, "grad_norm": 0.494140625, "learning_rate": 0.000491812875432127, "loss": 3.6242, "num_input_tokens_seen": 94371840, "step": 720, "train_runtime": 35767.721, "train_tokens_per_second": 2638.464 }, { "epoch": 0.444106463878327, "grad_norm": 0.62109375, "learning_rate": 0.0004914973849652431, "loss": 3.6366, "num_input_tokens_seen": 95682560, "step": 730, "train_runtime": 36285.9776, "train_tokens_per_second": 2636.902 }, { "epoch": 0.45019011406844106, "grad_norm": 0.474609375, "learning_rate": 0.0004911760352451052, "loss": 3.5994, "num_input_tokens_seen": 96993280, "step": 740, "train_runtime": 36794.874, "train_tokens_per_second": 2636.054 }, { "epoch": 0.45627376425855515, "grad_norm": 0.546875, "learning_rate": 0.0004908488340683583, "loss": 3.5819, "num_input_tokens_seen": 98304000, "step": 750, "train_runtime": 37301.9123, "train_tokens_per_second": 2635.361 }, { "epoch": 0.4623574144486692, "grad_norm": 0.51953125, "learning_rate": 0.0004905157893736169, "loss": 3.5663, "num_input_tokens_seen": 99614720, "step": 760, "train_runtime": 37810.1562, "train_tokens_per_second": 2634.602 }, { "epoch": 0.46844106463878327, "grad_norm": 0.482421875, "learning_rate": 0.0004901769092412711, "loss": 3.5701, "num_input_tokens_seen": 100925440, "step": 770, "train_runtime": 38317.6133, "train_tokens_per_second": 2633.918 }, { "epoch": 0.47452471482889735, "grad_norm": 0.5234375, "learning_rate": 0.0004898322018932924, "loss": 3.5661, "num_input_tokens_seen": 102236160, "step": 780, "train_runtime": 38820.0186, "train_tokens_per_second": 2633.594 }, { "epoch": 0.4806083650190114, "grad_norm": 0.52734375, "learning_rate": 0.0004894816756930327, "loss": 3.5268, "num_input_tokens_seen": 103546880, "step": 790, "train_runtime": 39328.8031, "train_tokens_per_second": 2632.851 }, { "epoch": 0.4866920152091255, "grad_norm": 0.486328125, "learning_rate": 0.0004891253391450222, "loss": 3.5552, "num_input_tokens_seen": 104857600, "step": 800, "train_runtime": 39836.4687, "train_tokens_per_second": 2632.201 }, { "epoch": 0.49277566539923956, "grad_norm": 0.41796875, "learning_rate": 0.0004887632008947625, "loss": 3.5054, "num_input_tokens_seen": 106168320, "step": 810, "train_runtime": 40344.8403, "train_tokens_per_second": 2631.522 }, { "epoch": 0.4988593155893536, "grad_norm": 0.490234375, "learning_rate": 0.0004883952697285176, "loss": 3.5364, "num_input_tokens_seen": 107479040, "step": 820, "train_runtime": 40853.9425, "train_tokens_per_second": 2630.812 }, { "epoch": 0.5049429657794677, "grad_norm": 0.416015625, "learning_rate": 0.0004880215545730999, "loss": 3.5194, "num_input_tokens_seen": 108789760, "step": 830, "train_runtime": 41371.0384, "train_tokens_per_second": 2629.612 }, { "epoch": 0.5110266159695818, "grad_norm": 0.427734375, "learning_rate": 0.00048764206449565414, "loss": 3.5011, "num_input_tokens_seen": 110100480, "step": 840, "train_runtime": 41878.8852, "train_tokens_per_second": 2629.021 }, { "epoch": 0.5171102661596958, "grad_norm": 0.5, "learning_rate": 0.00048725680870343726, "loss": 3.4902, "num_input_tokens_seen": 111411200, "step": 850, "train_runtime": 42387.2869, "train_tokens_per_second": 2628.411 }, { "epoch": 0.5231939163498099, "grad_norm": 0.458984375, "learning_rate": 0.000486865796543595, "loss": 3.4779, "num_input_tokens_seen": 112721920, "step": 860, "train_runtime": 42896.0857, "train_tokens_per_second": 2627.79 }, { "epoch": 0.529277566539924, "grad_norm": 0.51171875, "learning_rate": 0.0004864690375029351, "loss": 3.4634, "num_input_tokens_seen": 114032640, "step": 870, "train_runtime": 43408.6763, "train_tokens_per_second": 2626.955 }, { "epoch": 0.535361216730038, "grad_norm": 0.474609375, "learning_rate": 0.0004860665412076973, "loss": 3.4385, "num_input_tokens_seen": 115343360, "step": 880, "train_runtime": 43915.7309, "train_tokens_per_second": 2626.47 }, { "epoch": 0.5414448669201521, "grad_norm": 0.41796875, "learning_rate": 0.00048565831742331954, "loss": 3.4531, "num_input_tokens_seen": 116654080, "step": 890, "train_runtime": 44431.4678, "train_tokens_per_second": 2625.483 }, { "epoch": 0.5475285171102662, "grad_norm": 0.46484375, "learning_rate": 0.00048524437605420103, "loss": 3.4657, "num_input_tokens_seen": 117964800, "step": 900, "train_runtime": 44936.808, "train_tokens_per_second": 2625.126 }, { "epoch": 0.5536121673003802, "grad_norm": 0.52734375, "learning_rate": 0.00048482472714346227, "loss": 3.4579, "num_input_tokens_seen": 119275520, "step": 910, "train_runtime": 45444.6195, "train_tokens_per_second": 2624.635 }, { "epoch": 0.5596958174904944, "grad_norm": 0.4296875, "learning_rate": 0.000484399380872701, "loss": 3.4262, "num_input_tokens_seen": 120586240, "step": 920, "train_runtime": 45957.1389, "train_tokens_per_second": 2623.885 }, { "epoch": 0.5657794676806084, "grad_norm": 0.4375, "learning_rate": 0.0004839683475617455, "loss": 3.4126, "num_input_tokens_seen": 121896960, "step": 930, "train_runtime": 46465.2735, "train_tokens_per_second": 2623.399 }, { "epoch": 0.5718631178707224, "grad_norm": 0.80859375, "learning_rate": 0.00048353163766840385, "loss": 3.4239, "num_input_tokens_seen": 123207680, "step": 940, "train_runtime": 46973.782, "train_tokens_per_second": 2622.903 }, { "epoch": 0.5779467680608364, "grad_norm": 1.359375, "learning_rate": 0.00048308926178821054, "loss": 3.3839, "num_input_tokens_seen": 124518400, "step": 950, "train_runtime": 47481.0819, "train_tokens_per_second": 2622.484 }, { "epoch": 0.5840304182509506, "grad_norm": 1.4140625, "learning_rate": 0.00048264123065416934, "loss": 3.441, "num_input_tokens_seen": 125829120, "step": 960, "train_runtime": 47997.7462, "train_tokens_per_second": 2621.563 }, { "epoch": 0.5901140684410646, "grad_norm": 0.73046875, "learning_rate": 0.00048218755513649257, "loss": 3.4104, "num_input_tokens_seen": 127139840, "step": 970, "train_runtime": 48505.8936, "train_tokens_per_second": 2621.121 }, { "epoch": 0.5961977186311787, "grad_norm": 0.56640625, "learning_rate": 0.0004817282462423378, "loss": 3.3967, "num_input_tokens_seen": 128450560, "step": 980, "train_runtime": 49013.9882, "train_tokens_per_second": 2620.692 }, { "epoch": 0.6022813688212928, "grad_norm": 0.490234375, "learning_rate": 0.00048126331511554044, "loss": 3.4011, "num_input_tokens_seen": 129761280, "step": 990, "train_runtime": 49525.127, "train_tokens_per_second": 2620.11 }, { "epoch": 0.6083650190114068, "grad_norm": 0.419921875, "learning_rate": 0.0004807927730363437, "loss": 3.365, "num_input_tokens_seen": 131072000, "step": 1000, "train_runtime": 50032.5177, "train_tokens_per_second": 2619.736 }, { "epoch": 0.6144486692015209, "grad_norm": 0.51953125, "learning_rate": 0.00048031663142112464, "loss": 3.3622, "num_input_tokens_seen": 132382720, "step": 1010, "train_runtime": 50541.1055, "train_tokens_per_second": 2619.308 }, { "epoch": 0.620532319391635, "grad_norm": 1.2578125, "learning_rate": 0.0004798349018221173, "loss": 3.386, "num_input_tokens_seen": 133693440, "step": 1020, "train_runtime": 51052.0538, "train_tokens_per_second": 2618.767 }, { "epoch": 0.626615969581749, "grad_norm": 3.5625, "learning_rate": 0.0004793475959271323, "loss": 3.3605, "num_input_tokens_seen": 135004160, "step": 1030, "train_runtime": 51561.7454, "train_tokens_per_second": 2618.301 }, { "epoch": 0.6326996197718631, "grad_norm": 0.609375, "learning_rate": 0.0004788547255592736, "loss": 3.3946, "num_input_tokens_seen": 136314880, "step": 1040, "train_runtime": 52071.6302, "train_tokens_per_second": 2617.834 }, { "epoch": 0.6387832699619772, "grad_norm": 0.5234375, "learning_rate": 0.00047835630267665114, "loss": 3.3573, "num_input_tokens_seen": 137625600, "step": 1050, "train_runtime": 52586.3476, "train_tokens_per_second": 2617.136 }, { "epoch": 0.6448669201520912, "grad_norm": 0.408203125, "learning_rate": 0.0004778523393720911, "loss": 3.3555, "num_input_tokens_seen": 138936320, "step": 1060, "train_runtime": 53105.6264, "train_tokens_per_second": 2616.226 }, { "epoch": 0.6509505703422053, "grad_norm": 0.447265625, "learning_rate": 0.00047734284787284237, "loss": 3.3005, "num_input_tokens_seen": 140247040, "step": 1070, "train_runtime": 53616.3439, "train_tokens_per_second": 2615.752 }, { "epoch": 0.6570342205323194, "grad_norm": 0.44140625, "learning_rate": 0.00047682784054027994, "loss": 3.3505, "num_input_tokens_seen": 141557760, "step": 1080, "train_runtime": 54124.0806, "train_tokens_per_second": 2615.43 }, { "epoch": 0.6631178707224334, "grad_norm": 0.38671875, "learning_rate": 0.00047630732986960486, "loss": 3.3432, "num_input_tokens_seen": 142868480, "step": 1090, "train_runtime": 54632.5779, "train_tokens_per_second": 2615.079 }, { "epoch": 0.6692015209125475, "grad_norm": 0.447265625, "learning_rate": 0.0004757813284895412, "loss": 3.329, "num_input_tokens_seen": 144179200, "step": 1100, "train_runtime": 55142.8421, "train_tokens_per_second": 2614.649 }, { "epoch": 0.6752851711026616, "grad_norm": 0.453125, "learning_rate": 0.0004752498491620296, "loss": 3.3267, "num_input_tokens_seen": 145489920, "step": 1110, "train_runtime": 55654.8916, "train_tokens_per_second": 2614.144 }, { "epoch": 0.6813688212927757, "grad_norm": 0.462890625, "learning_rate": 0.00047471290478191786, "loss": 3.3118, "num_input_tokens_seen": 146800640, "step": 1120, "train_runtime": 56162.9248, "train_tokens_per_second": 2613.835 }, { "epoch": 0.6874524714828897, "grad_norm": 0.62890625, "learning_rate": 0.0004741705083766475, "loss": 3.357, "num_input_tokens_seen": 148111360, "step": 1130, "train_runtime": 56673.7299, "train_tokens_per_second": 2613.404 }, { "epoch": 0.6935361216730038, "grad_norm": 0.55078125, "learning_rate": 0.0004736226731059383, "loss": 3.3046, "num_input_tokens_seen": 149422080, "step": 1140, "train_runtime": 57182.6726, "train_tokens_per_second": 2613.066 }, { "epoch": 0.6996197718631179, "grad_norm": 0.443359375, "learning_rate": 0.00047306941226146884, "loss": 3.3162, "num_input_tokens_seen": 150732800, "step": 1150, "train_runtime": 57691.3075, "train_tokens_per_second": 2612.747 }, { "epoch": 0.7057034220532319, "grad_norm": 0.486328125, "learning_rate": 0.0004725107392665536, "loss": 3.2922, "num_input_tokens_seen": 152043520, "step": 1160, "train_runtime": 58195.4974, "train_tokens_per_second": 2612.634 }, { "epoch": 0.711787072243346, "grad_norm": 0.451171875, "learning_rate": 0.0004719466676758181, "loss": 3.3052, "num_input_tokens_seen": 153354240, "step": 1170, "train_runtime": 58702.65, "train_tokens_per_second": 2612.39 }, { "epoch": 0.7178707224334601, "grad_norm": 0.46484375, "learning_rate": 0.0004713772111748693, "loss": 3.3115, "num_input_tokens_seen": 154664960, "step": 1180, "train_runtime": 59213.6331, "train_tokens_per_second": 2611.982 }, { "epoch": 0.7239543726235741, "grad_norm": 0.416015625, "learning_rate": 0.0004708023835799638, "loss": 3.2765, "num_input_tokens_seen": 155975680, "step": 1190, "train_runtime": 59725.0865, "train_tokens_per_second": 2611.561 }, { "epoch": 0.7300380228136882, "grad_norm": 0.40234375, "learning_rate": 0.0004702221988376728, "loss": 3.2856, "num_input_tokens_seen": 157286400, "step": 1200, "train_runtime": 60234.323, "train_tokens_per_second": 2611.242 }, { "epoch": 0.7361216730038023, "grad_norm": 0.390625, "learning_rate": 0.00046963667102454344, "loss": 3.2904, "num_input_tokens_seen": 158597120, "step": 1210, "train_runtime": 60742.9574, "train_tokens_per_second": 2610.955 }, { "epoch": 0.7422053231939163, "grad_norm": 0.70703125, "learning_rate": 0.00046904581434675753, "loss": 3.2525, "num_input_tokens_seen": 159907840, "step": 1220, "train_runtime": 61259.4126, "train_tokens_per_second": 2610.339 }, { "epoch": 0.7482889733840304, "grad_norm": 0.52734375, "learning_rate": 0.00046844964313978667, "loss": 3.2743, "num_input_tokens_seen": 161218560, "step": 1230, "train_runtime": 61770.0765, "train_tokens_per_second": 2609.978 }, { "epoch": 0.7543726235741445, "grad_norm": 0.443359375, "learning_rate": 0.00046784817186804463, "loss": 3.2488, "num_input_tokens_seen": 162529280, "step": 1240, "train_runtime": 62278.1018, "train_tokens_per_second": 2609.734 }, { "epoch": 0.7604562737642585, "grad_norm": 0.443359375, "learning_rate": 0.0004672414151245361, "loss": 3.2617, "num_input_tokens_seen": 163840000, "step": 1250, "train_runtime": 62782.4473, "train_tokens_per_second": 2609.647 }, { "epoch": 0.7665399239543726, "grad_norm": 0.42578125, "learning_rate": 0.0004666293876305031, "loss": 3.2405, "num_input_tokens_seen": 165150720, "step": 1260, "train_runtime": 63294.8295, "train_tokens_per_second": 2609.229 }, { "epoch": 0.7726235741444867, "grad_norm": 0.36328125, "learning_rate": 0.00046601210423506743, "loss": 3.2632, "num_input_tokens_seen": 166461440, "step": 1270, "train_runtime": 63800.2584, "train_tokens_per_second": 2609.103 }, { "epoch": 0.7787072243346007, "grad_norm": 0.40625, "learning_rate": 0.0004653895799148706, "loss": 3.2089, "num_input_tokens_seen": 167772160, "step": 1280, "train_runtime": 64315.6024, "train_tokens_per_second": 2608.576 }, { "epoch": 0.7847908745247149, "grad_norm": 0.419921875, "learning_rate": 0.00046476182977371014, "loss": 3.2395, "num_input_tokens_seen": 169082880, "step": 1290, "train_runtime": 64830.3388, "train_tokens_per_second": 2608.083 }, { "epoch": 0.7908745247148289, "grad_norm": 0.38671875, "learning_rate": 0.0004641288690421737, "loss": 3.2493, "num_input_tokens_seen": 170393600, "step": 1300, "train_runtime": 65347.315, "train_tokens_per_second": 2607.507 }, { "epoch": 0.7969581749049429, "grad_norm": 0.365234375, "learning_rate": 0.000463490713077269, "loss": 3.2235, "num_input_tokens_seen": 171704320, "step": 1310, "train_runtime": 65855.9256, "train_tokens_per_second": 2607.272 }, { "epoch": 0.8030418250950571, "grad_norm": 0.41015625, "learning_rate": 0.00046284737736205155, "loss": 3.2458, "num_input_tokens_seen": 173015040, "step": 1320, "train_runtime": 66363.9708, "train_tokens_per_second": 2607.063 }, { "epoch": 0.8091254752851711, "grad_norm": 0.392578125, "learning_rate": 0.00046219887750524895, "loss": 3.2229, "num_input_tokens_seen": 174325760, "step": 1330, "train_runtime": 66872.6674, "train_tokens_per_second": 2606.831 }, { "epoch": 0.8152091254752851, "grad_norm": 0.37890625, "learning_rate": 0.00046154522924088203, "loss": 3.2096, "num_input_tokens_seen": 175636480, "step": 1340, "train_runtime": 67386.1096, "train_tokens_per_second": 2606.42 }, { "epoch": 0.8212927756653993, "grad_norm": 0.369140625, "learning_rate": 0.00046088644842788335, "loss": 3.1919, "num_input_tokens_seen": 176947200, "step": 1350, "train_runtime": 67893.1543, "train_tokens_per_second": 2606.26 }, { "epoch": 0.8273764258555133, "grad_norm": 0.42578125, "learning_rate": 0.0004602225510497121, "loss": 3.2157, "num_input_tokens_seen": 178257920, "step": 1360, "train_runtime": 68401.5702, "train_tokens_per_second": 2606.05 }, { "epoch": 0.8334600760456273, "grad_norm": 0.3984375, "learning_rate": 0.00045955355321396663, "loss": 3.1801, "num_input_tokens_seen": 179568640, "step": 1370, "train_runtime": 68913.0159, "train_tokens_per_second": 2605.729 }, { "epoch": 0.8395437262357415, "grad_norm": 0.365234375, "learning_rate": 0.0004588794711519935, "loss": 3.1953, "num_input_tokens_seen": 180879360, "step": 1380, "train_runtime": 69424.3103, "train_tokens_per_second": 2605.418 }, { "epoch": 0.8456273764258555, "grad_norm": 0.357421875, "learning_rate": 0.00045820032121849367, "loss": 3.2086, "num_input_tokens_seen": 182190080, "step": 1390, "train_runtime": 69934.3095, "train_tokens_per_second": 2605.16 }, { "epoch": 0.8517110266159695, "grad_norm": 0.416015625, "learning_rate": 0.0004575161198911256, "loss": 3.1908, "num_input_tokens_seen": 183500800, "step": 1400, "train_runtime": 70445.1369, "train_tokens_per_second": 2604.875 }, { "epoch": 0.8577946768060837, "grad_norm": 0.392578125, "learning_rate": 0.00045682688377010595, "loss": 3.193, "num_input_tokens_seen": 184811520, "step": 1410, "train_runtime": 70955.9519, "train_tokens_per_second": 2604.595 }, { "epoch": 0.8638783269961977, "grad_norm": 0.373046875, "learning_rate": 0.000456132629577806, "loss": 3.1847, "num_input_tokens_seen": 186122240, "step": 1420, "train_runtime": 71465.9376, "train_tokens_per_second": 2604.349 }, { "epoch": 0.8699619771863117, "grad_norm": 0.3671875, "learning_rate": 0.00045543337415834676, "loss": 3.1901, "num_input_tokens_seen": 187432960, "step": 1430, "train_runtime": 71976.5411, "train_tokens_per_second": 2604.084 }, { "epoch": 0.8760456273764259, "grad_norm": 0.39453125, "learning_rate": 0.0004547291344771896, "loss": 3.2061, "num_input_tokens_seen": 188743680, "step": 1440, "train_runtime": 72494.8149, "train_tokens_per_second": 2603.547 }, { "epoch": 0.8821292775665399, "grad_norm": 0.36328125, "learning_rate": 0.0004540199276207252, "loss": 3.2118, "num_input_tokens_seen": 190054400, "step": 1450, "train_runtime": 73006.8923, "train_tokens_per_second": 2603.239 }, { "epoch": 0.888212927756654, "grad_norm": 0.369140625, "learning_rate": 0.00045330577079585855, "loss": 3.1828, "num_input_tokens_seen": 191365120, "step": 1460, "train_runtime": 73514.9522, "train_tokens_per_second": 2603.078 }, { "epoch": 0.8942965779467681, "grad_norm": 0.427734375, "learning_rate": 0.0004525866813295919, "loss": 3.2098, "num_input_tokens_seen": 192675840, "step": 1470, "train_runtime": 74022.1124, "train_tokens_per_second": 2602.95 }, { "epoch": 0.9003802281368821, "grad_norm": 0.359375, "learning_rate": 0.00045186267666860394, "loss": 3.176, "num_input_tokens_seen": 193986560, "step": 1480, "train_runtime": 74530.9401, "train_tokens_per_second": 2602.766 }, { "epoch": 0.9064638783269962, "grad_norm": 0.357421875, "learning_rate": 0.0004511337743788266, "loss": 3.1603, "num_input_tokens_seen": 195297280, "step": 1490, "train_runtime": 75039.6371, "train_tokens_per_second": 2602.588 }, { "epoch": 0.9125475285171103, "grad_norm": 0.3671875, "learning_rate": 0.00045039999214501944, "loss": 3.1512, "num_input_tokens_seen": 196608000, "step": 1500, "train_runtime": 75560.5392, "train_tokens_per_second": 2601.993 }, { "epoch": 0.9186311787072243, "grad_norm": 0.384765625, "learning_rate": 0.00044966134777033956, "loss": 3.1639, "num_input_tokens_seen": 197918720, "step": 1510, "train_runtime": 76079.4478, "train_tokens_per_second": 2601.474 }, { "epoch": 0.9247148288973384, "grad_norm": 0.361328125, "learning_rate": 0.00044891785917591055, "loss": 3.1389, "num_input_tokens_seen": 199229440, "step": 1520, "train_runtime": 76592.6237, "train_tokens_per_second": 2601.157 }, { "epoch": 0.9307984790874525, "grad_norm": 0.38671875, "learning_rate": 0.0004481695444003871, "loss": 3.1475, "num_input_tokens_seen": 200540160, "step": 1530, "train_runtime": 77104.1479, "train_tokens_per_second": 2600.9 }, { "epoch": 0.9368821292775665, "grad_norm": 0.421875, "learning_rate": 0.00044741642159951795, "loss": 3.1367, "num_input_tokens_seen": 201850880, "step": 1540, "train_runtime": 77615.4676, "train_tokens_per_second": 2600.653 }, { "epoch": 0.9429657794676806, "grad_norm": 0.373046875, "learning_rate": 0.0004466585090457046, "loss": 3.1162, "num_input_tokens_seen": 203161600, "step": 1550, "train_runtime": 78127.8408, "train_tokens_per_second": 2600.374 }, { "epoch": 0.9490494296577947, "grad_norm": 0.345703125, "learning_rate": 0.00044589582512755854, "loss": 3.136, "num_input_tokens_seen": 204472320, "step": 1560, "train_runtime": 78643.2833, "train_tokens_per_second": 2599.997 }, { "epoch": 0.9551330798479087, "grad_norm": 0.404296875, "learning_rate": 0.0004451283883494551, "loss": 3.1143, "num_input_tokens_seen": 205783040, "step": 1570, "train_runtime": 79162.5028, "train_tokens_per_second": 2599.501 }, { "epoch": 0.9612167300380228, "grad_norm": 0.35546875, "learning_rate": 0.0004443562173310842, "loss": 3.1332, "num_input_tokens_seen": 207093760, "step": 1580, "train_runtime": 79673.0445, "train_tokens_per_second": 2599.295 }, { "epoch": 0.9673003802281369, "grad_norm": 0.38671875, "learning_rate": 0.00044357933080699886, "loss": 3.1363, "num_input_tokens_seen": 208404480, "step": 1590, "train_runtime": 80187.555, "train_tokens_per_second": 2598.963 }, { "epoch": 0.973384030418251, "grad_norm": 0.365234375, "learning_rate": 0.00044279774762616056, "loss": 3.1414, "num_input_tokens_seen": 209715200, "step": 1600, "train_runtime": 80700.7756, "train_tokens_per_second": 2598.676 }, { "epoch": 0.979467680608365, "grad_norm": 0.353515625, "learning_rate": 0.00044201148675148175, "loss": 3.1076, "num_input_tokens_seen": 211025920, "step": 1610, "train_runtime": 81218.0495, "train_tokens_per_second": 2598.264 }, { "epoch": 0.9855513307984791, "grad_norm": 0.365234375, "learning_rate": 0.000441220567259366, "loss": 3.1153, "num_input_tokens_seen": 212336640, "step": 1620, "train_runtime": 81730.0271, "train_tokens_per_second": 2598.025 }, { "epoch": 0.9916349809885932, "grad_norm": 0.380859375, "learning_rate": 0.00044042500833924535, "loss": 3.1097, "num_input_tokens_seen": 213647360, "step": 1630, "train_runtime": 82244.6699, "train_tokens_per_second": 2597.705 }, { "epoch": 0.9977186311787072, "grad_norm": 0.453125, "learning_rate": 0.0004396248292931141, "loss": 3.0839, "num_input_tokens_seen": 214958080, "step": 1640, "train_runtime": 82763.7003, "train_tokens_per_second": 2597.251 }, { "epoch": 1.0036501901140684, "grad_norm": 0.408203125, "learning_rate": 0.0004388200495350612, "loss": 2.9969, "num_input_tokens_seen": 216227840, "step": 1650, "train_runtime": 83205.3748, "train_tokens_per_second": 2598.724 }, { "epoch": 1.0097338403041825, "grad_norm": 0.41796875, "learning_rate": 0.0004380106885907987, "loss": 2.9191, "num_input_tokens_seen": 217538560, "step": 1660, "train_runtime": 83652.3328, "train_tokens_per_second": 2600.508 }, { "epoch": 1.0158174904942965, "grad_norm": 0.40234375, "learning_rate": 0.0004371967660971883, "loss": 2.897, "num_input_tokens_seen": 218849280, "step": 1670, "train_runtime": 84099.6608, "train_tokens_per_second": 2602.261 }, { "epoch": 1.0219011406844107, "grad_norm": 0.419921875, "learning_rate": 0.00043637830180176475, "loss": 2.8774, "num_input_tokens_seen": 220160000, "step": 1680, "train_runtime": 84541.9367, "train_tokens_per_second": 2604.151 }, { "epoch": 1.0279847908745248, "grad_norm": 0.41796875, "learning_rate": 0.00043555531556225695, "loss": 2.9148, "num_input_tokens_seen": 221470720, "step": 1690, "train_runtime": 84990.5402, "train_tokens_per_second": 2605.828 }, { "epoch": 1.0340684410646388, "grad_norm": 0.39453125, "learning_rate": 0.00043472782734610605, "loss": 2.8941, "num_input_tokens_seen": 222781440, "step": 1700, "train_runtime": 85437.717, "train_tokens_per_second": 2607.53 }, { "epoch": 1.0401520912547528, "grad_norm": 1.0234375, "learning_rate": 0.0004338958572299807, "loss": 2.9016, "num_input_tokens_seen": 224092160, "step": 1710, "train_runtime": 85885.9521, "train_tokens_per_second": 2609.183 }, { "epoch": 1.0462357414448669, "grad_norm": 0.47265625, "learning_rate": 0.00043305942539929057, "loss": 2.904, "num_input_tokens_seen": 225402880, "step": 1720, "train_runtime": 86334.27, "train_tokens_per_second": 2610.816 }, { "epoch": 1.052319391634981, "grad_norm": 0.44140625, "learning_rate": 0.00043221855214769606, "loss": 2.9289, "num_input_tokens_seen": 226713600, "step": 1730, "train_runtime": 86781.5406, "train_tokens_per_second": 2612.463 }, { "epoch": 1.0584030418250951, "grad_norm": 0.40625, "learning_rate": 0.0004313732578766161, "loss": 2.9148, "num_input_tokens_seen": 228024320, "step": 1740, "train_runtime": 87228.9311, "train_tokens_per_second": 2614.09 }, { "epoch": 1.0644866920152092, "grad_norm": 0.41015625, "learning_rate": 0.00043052356309473326, "loss": 2.9057, "num_input_tokens_seen": 229335040, "step": 1750, "train_runtime": 87675.2394, "train_tokens_per_second": 2615.733 }, { "epoch": 1.0705703422053232, "grad_norm": 0.380859375, "learning_rate": 0.0004296694884174962, "loss": 2.9202, "num_input_tokens_seen": 230645760, "step": 1760, "train_runtime": 88121.946, "train_tokens_per_second": 2617.348 }, { "epoch": 1.0766539923954372, "grad_norm": 0.408203125, "learning_rate": 0.00042881105456661915, "loss": 2.9456, "num_input_tokens_seen": 231956480, "step": 1770, "train_runtime": 88570.7729, "train_tokens_per_second": 2618.883 }, { "epoch": 1.0827376425855513, "grad_norm": 0.390625, "learning_rate": 0.0004279482823695797, "loss": 2.9425, "num_input_tokens_seen": 233267200, "step": 1780, "train_runtime": 89019.7651, "train_tokens_per_second": 2620.398 }, { "epoch": 1.0888212927756653, "grad_norm": 0.4375, "learning_rate": 0.00042708119275911276, "loss": 2.9194, "num_input_tokens_seen": 234577920, "step": 1790, "train_runtime": 89469.0148, "train_tokens_per_second": 2621.89 }, { "epoch": 1.0949049429657796, "grad_norm": 0.41015625, "learning_rate": 0.0004262098067727036, "loss": 2.8975, "num_input_tokens_seen": 235888640, "step": 1800, "train_runtime": 89917.0004, "train_tokens_per_second": 2623.404 }, { "epoch": 1.1009885931558936, "grad_norm": 0.35546875, "learning_rate": 0.0004253341455520766, "loss": 2.9038, "num_input_tokens_seen": 237199360, "step": 1810, "train_runtime": 90363.5211, "train_tokens_per_second": 2624.946 }, { "epoch": 1.1070722433460076, "grad_norm": 0.365234375, "learning_rate": 0.0004244542303426828, "loss": 2.9133, "num_input_tokens_seen": 238510080, "step": 1820, "train_runtime": 90809.803, "train_tokens_per_second": 2626.479 }, { "epoch": 1.1131558935361217, "grad_norm": 0.37890625, "learning_rate": 0.0004235700824931841, "loss": 2.9203, "num_input_tokens_seen": 239820800, "step": 1830, "train_runtime": 91256.862, "train_tokens_per_second": 2627.976 }, { "epoch": 1.1192395437262357, "grad_norm": 0.369140625, "learning_rate": 0.0004226817234549358, "loss": 2.8757, "num_input_tokens_seen": 241131520, "step": 1840, "train_runtime": 91704.3196, "train_tokens_per_second": 2629.446 }, { "epoch": 1.1253231939163497, "grad_norm": 0.39453125, "learning_rate": 0.0004217891747814656, "loss": 2.915, "num_input_tokens_seen": 242442240, "step": 1850, "train_runtime": 92152.2557, "train_tokens_per_second": 2630.888 }, { "epoch": 1.131406844106464, "grad_norm": 0.435546875, "learning_rate": 0.00042089245812795085, "loss": 2.9151, "num_input_tokens_seen": 243752960, "step": 1860, "train_runtime": 92600.8297, "train_tokens_per_second": 2632.298 }, { "epoch": 1.137490494296578, "grad_norm": 5.59375, "learning_rate": 0.0004199915952506933, "loss": 2.8949, "num_input_tokens_seen": 245063680, "step": 1870, "train_runtime": 93047.388, "train_tokens_per_second": 2633.751 }, { "epoch": 1.143574144486692, "grad_norm": 0.6171875, "learning_rate": 0.00041908660800659103, "loss": 2.9332, "num_input_tokens_seen": 246374400, "step": 1880, "train_runtime": 93494.6982, "train_tokens_per_second": 2635.17 }, { "epoch": 1.149657794676806, "grad_norm": 0.6328125, "learning_rate": 0.00041817751835260844, "loss": 2.9199, "num_input_tokens_seen": 247685120, "step": 1890, "train_runtime": 93942.5611, "train_tokens_per_second": 2636.559 }, { "epoch": 1.15574144486692, "grad_norm": 0.380859375, "learning_rate": 0.0004172643483452431, "loss": 2.9073, "num_input_tokens_seen": 248995840, "step": 1900, "train_runtime": 94390.1139, "train_tokens_per_second": 2637.944 }, { "epoch": 1.1618250950570341, "grad_norm": 1.3984375, "learning_rate": 0.00041634712013999107, "loss": 2.9329, "num_input_tokens_seen": 250306560, "step": 1910, "train_runtime": 94837.9896, "train_tokens_per_second": 2639.307 }, { "epoch": 1.1679087452471484, "grad_norm": 0.59375, "learning_rate": 0.00041542585599080897, "loss": 2.8958, "num_input_tokens_seen": 251617280, "step": 1920, "train_runtime": 95285.7174, "train_tokens_per_second": 2640.661 }, { "epoch": 1.1739923954372624, "grad_norm": 57.75, "learning_rate": 0.0004145005782495743, "loss": 2.9621, "num_input_tokens_seen": 252928000, "step": 1930, "train_runtime": 95734.3074, "train_tokens_per_second": 2641.979 }, { "epoch": 1.1800760456273764, "grad_norm": 4.25, "learning_rate": 0.0004135713093655431, "loss": 2.9531, "num_input_tokens_seen": 254238720, "step": 1940, "train_runtime": 96179.9871, "train_tokens_per_second": 2643.364 }, { "epoch": 1.1861596958174905, "grad_norm": 2.46875, "learning_rate": 0.0004126380718848052, "loss": 2.9254, "num_input_tokens_seen": 255549440, "step": 1950, "train_runtime": 96629.2522, "train_tokens_per_second": 2644.638 }, { "epoch": 1.1922433460076045, "grad_norm": 0.77734375, "learning_rate": 0.00041170088844973715, "loss": 2.9068, "num_input_tokens_seen": 256860160, "step": 1960, "train_runtime": 97084.0321, "train_tokens_per_second": 2645.751 }, { "epoch": 1.1983269961977185, "grad_norm": 0.55859375, "learning_rate": 0.0004107597817984531, "loss": 2.9029, "num_input_tokens_seen": 258170880, "step": 1970, "train_runtime": 97539.64, "train_tokens_per_second": 2646.83 }, { "epoch": 1.2044106463878328, "grad_norm": 0.421875, "learning_rate": 0.0004098147747642529, "loss": 2.892, "num_input_tokens_seen": 259481600, "step": 1980, "train_runtime": 97994.7012, "train_tokens_per_second": 2647.915 }, { "epoch": 1.2104942965779468, "grad_norm": 0.462890625, "learning_rate": 0.00040886589027506814, "loss": 2.8836, "num_input_tokens_seen": 260792320, "step": 1990, "train_runtime": 98450.2354, "train_tokens_per_second": 2648.976 }, { "epoch": 1.2165779467680609, "grad_norm": 0.466796875, "learning_rate": 0.0004079131513529062, "loss": 2.897, "num_input_tokens_seen": 262103040, "step": 2000, "train_runtime": 98899.4674, "train_tokens_per_second": 2650.197 }, { "epoch": 1.222661596958175, "grad_norm": 0.490234375, "learning_rate": 0.00040695658111329104, "loss": 2.8802, "num_input_tokens_seen": 263413760, "step": 2010, "train_runtime": 99347.728, "train_tokens_per_second": 2651.432 }, { "epoch": 1.228745247148289, "grad_norm": 0.5078125, "learning_rate": 0.00040599620276470297, "loss": 2.8942, "num_input_tokens_seen": 264724480, "step": 2020, "train_runtime": 99795.6938, "train_tokens_per_second": 2652.664 }, { "epoch": 1.234828897338403, "grad_norm": 0.466796875, "learning_rate": 0.00040503203960801516, "loss": 2.8637, "num_input_tokens_seen": 266035200, "step": 2030, "train_runtime": 100246.5331, "train_tokens_per_second": 2653.809 }, { "epoch": 1.2409125475285172, "grad_norm": 0.427734375, "learning_rate": 0.0004040641150359288, "loss": 2.8851, "num_input_tokens_seen": 267345920, "step": 2040, "train_runtime": 100696.8528, "train_tokens_per_second": 2654.958 }, { "epoch": 1.2469961977186312, "grad_norm": 0.447265625, "learning_rate": 0.0004030924525324049, "loss": 2.8836, "num_input_tokens_seen": 268656640, "step": 2050, "train_runtime": 101148.4193, "train_tokens_per_second": 2656.064 }, { "epoch": 1.2530798479087453, "grad_norm": 0.41015625, "learning_rate": 0.0004021170756720949, "loss": 2.8915, "num_input_tokens_seen": 269967360, "step": 2060, "train_runtime": 101597.8838, "train_tokens_per_second": 2657.214 }, { "epoch": 1.2591634980988593, "grad_norm": 0.412109375, "learning_rate": 0.0004011380081197687, "loss": 2.9017, "num_input_tokens_seen": 271278080, "step": 2070, "train_runtime": 102048.5088, "train_tokens_per_second": 2658.325 }, { "epoch": 1.2652471482889733, "grad_norm": 0.41015625, "learning_rate": 0.00040015527362974047, "loss": 2.8829, "num_input_tokens_seen": 272588800, "step": 2080, "train_runtime": 102496.6301, "train_tokens_per_second": 2659.49 }, { "epoch": 1.2713307984790876, "grad_norm": 0.369140625, "learning_rate": 0.00039916889604529215, "loss": 2.8651, "num_input_tokens_seen": 273899520, "step": 2090, "train_runtime": 102945.3454, "train_tokens_per_second": 2660.63 }, { "epoch": 1.2774144486692016, "grad_norm": 0.365234375, "learning_rate": 0.00039817889929809536, "loss": 2.9117, "num_input_tokens_seen": 275210240, "step": 2100, "train_runtime": 103394.12, "train_tokens_per_second": 2661.759 }, { "epoch": 1.2834980988593157, "grad_norm": 0.435546875, "learning_rate": 0.00039718530740763015, "loss": 2.8734, "num_input_tokens_seen": 276520960, "step": 2110, "train_runtime": 103841.6282, "train_tokens_per_second": 2662.91 }, { "epoch": 1.2895817490494297, "grad_norm": 0.3984375, "learning_rate": 0.0003961881444806028, "loss": 2.8978, "num_input_tokens_seen": 277831680, "step": 2120, "train_runtime": 104294.6414, "train_tokens_per_second": 2663.911 }, { "epoch": 1.2956653992395437, "grad_norm": 0.36328125, "learning_rate": 0.0003951874347103607, "loss": 2.8918, "num_input_tokens_seen": 279142400, "step": 2130, "train_runtime": 104743.1678, "train_tokens_per_second": 2665.018 }, { "epoch": 1.3017490494296577, "grad_norm": 0.375, "learning_rate": 0.00039418320237630543, "loss": 2.8742, "num_input_tokens_seen": 280453120, "step": 2140, "train_runtime": 105191.5091, "train_tokens_per_second": 2666.119 }, { "epoch": 1.3078326996197718, "grad_norm": 0.373046875, "learning_rate": 0.00039317547184330347, "loss": 2.8766, "num_input_tokens_seen": 281763840, "step": 2150, "train_runtime": 105640.1283, "train_tokens_per_second": 2667.205 }, { "epoch": 1.3139163498098858, "grad_norm": 0.349609375, "learning_rate": 0.0003921642675610956, "loss": 2.8618, "num_input_tokens_seen": 283074560, "step": 2160, "train_runtime": 106090.5411, "train_tokens_per_second": 2668.236 }, { "epoch": 1.32, "grad_norm": 0.427734375, "learning_rate": 0.0003911496140637028, "loss": 2.9176, "num_input_tokens_seen": 284385280, "step": 2170, "train_runtime": 106543.4269, "train_tokens_per_second": 2669.196 }, { "epoch": 1.326083650190114, "grad_norm": 0.3359375, "learning_rate": 0.00039013153596883203, "loss": 2.8832, "num_input_tokens_seen": 285696000, "step": 2180, "train_runtime": 106996.4598, "train_tokens_per_second": 2670.144 }, { "epoch": 1.3321673003802281, "grad_norm": 0.396484375, "learning_rate": 0.00038911005797727816, "loss": 2.8554, "num_input_tokens_seen": 287006720, "step": 2190, "train_runtime": 107450.6455, "train_tokens_per_second": 2671.056 }, { "epoch": 1.3382509505703422, "grad_norm": 0.3984375, "learning_rate": 0.00038808520487232514, "loss": 2.9136, "num_input_tokens_seen": 288317440, "step": 2200, "train_runtime": 107900.1405, "train_tokens_per_second": 2672.077 }, { "epoch": 1.3443346007604564, "grad_norm": 0.36328125, "learning_rate": 0.00038705700151914446, "loss": 2.9164, "num_input_tokens_seen": 289628160, "step": 2210, "train_runtime": 108349.4576, "train_tokens_per_second": 2673.093 }, { "epoch": 1.3504182509505704, "grad_norm": 0.390625, "learning_rate": 0.0003860254728641918, "loss": 2.8747, "num_input_tokens_seen": 290938880, "step": 2220, "train_runtime": 108798.3026, "train_tokens_per_second": 2674.112 }, { "epoch": 1.3565019011406845, "grad_norm": 0.3828125, "learning_rate": 0.00038499064393460213, "loss": 2.8488, "num_input_tokens_seen": 292249600, "step": 2230, "train_runtime": 109248.832, "train_tokens_per_second": 2675.082 }, { "epoch": 1.3625855513307985, "grad_norm": 0.357421875, "learning_rate": 0.00038395253983758226, "loss": 2.8678, "num_input_tokens_seen": 293560320, "step": 2240, "train_runtime": 109700.3985, "train_tokens_per_second": 2676.019 }, { "epoch": 1.3686692015209125, "grad_norm": 0.4140625, "learning_rate": 0.00038291118575980147, "loss": 2.8556, "num_input_tokens_seen": 294871040, "step": 2250, "train_runtime": 110151.5807, "train_tokens_per_second": 2676.957 }, { "epoch": 1.3747528517110266, "grad_norm": 0.373046875, "learning_rate": 0.0003818666069667811, "loss": 2.8687, "num_input_tokens_seen": 296181760, "step": 2260, "train_runtime": 110602.5281, "train_tokens_per_second": 2677.893 }, { "epoch": 1.3808365019011406, "grad_norm": 0.36328125, "learning_rate": 0.0003808188288022806, "loss": 2.8703, "num_input_tokens_seen": 297492480, "step": 2270, "train_runtime": 111055.5295, "train_tokens_per_second": 2678.772 }, { "epoch": 1.3869201520912546, "grad_norm": 0.373046875, "learning_rate": 0.00037976787668768353, "loss": 2.8479, "num_input_tokens_seen": 298803200, "step": 2280, "train_runtime": 111506.098, "train_tokens_per_second": 2679.703 }, { "epoch": 1.3930038022813689, "grad_norm": 0.357421875, "learning_rate": 0.00037871377612138014, "loss": 2.8656, "num_input_tokens_seen": 300113920, "step": 2290, "train_runtime": 111958.2058, "train_tokens_per_second": 2680.589 }, { "epoch": 1.399087452471483, "grad_norm": 0.37890625, "learning_rate": 0.00037765655267814926, "loss": 2.854, "num_input_tokens_seen": 301424640, "step": 2300, "train_runtime": 112412.2892, "train_tokens_per_second": 2681.421 }, { "epoch": 1.405171102661597, "grad_norm": 0.37890625, "learning_rate": 0.0003765962320085373, "loss": 2.8524, "num_input_tokens_seen": 302735360, "step": 2310, "train_runtime": 112865.0758, "train_tokens_per_second": 2682.277 }, { "epoch": 1.411254752851711, "grad_norm": 0.375, "learning_rate": 0.00037553283983823615, "loss": 2.8348, "num_input_tokens_seen": 304046080, "step": 2320, "train_runtime": 113318.0386, "train_tokens_per_second": 2683.122 }, { "epoch": 1.417338403041825, "grad_norm": 0.359375, "learning_rate": 0.00037446640196745927, "loss": 2.8776, "num_input_tokens_seen": 305356800, "step": 2330, "train_runtime": 113769.7765, "train_tokens_per_second": 2683.989 }, { "epoch": 1.4234220532319393, "grad_norm": 0.361328125, "learning_rate": 0.0003733969442703151, "loss": 2.8572, "num_input_tokens_seen": 306667520, "step": 2340, "train_runtime": 114222.6283, "train_tokens_per_second": 2684.823 }, { "epoch": 1.4295057034220533, "grad_norm": 0.380859375, "learning_rate": 0.00037232449269417983, "loss": 2.8366, "num_input_tokens_seen": 307978240, "step": 2350, "train_runtime": 114676.9671, "train_tokens_per_second": 2685.615 }, { "epoch": 1.4355893536121673, "grad_norm": 0.365234375, "learning_rate": 0.0003712490732590677, "loss": 2.837, "num_input_tokens_seen": 309288960, "step": 2360, "train_runtime": 115129.6387, "train_tokens_per_second": 2686.441 }, { "epoch": 1.4416730038022814, "grad_norm": 0.361328125, "learning_rate": 0.0003701707120569997, "loss": 2.8182, "num_input_tokens_seen": 310599680, "step": 2370, "train_runtime": 115583.0774, "train_tokens_per_second": 2687.242 }, { "epoch": 1.4477566539923954, "grad_norm": 0.33984375, "learning_rate": 0.0003690894352513703, "loss": 2.8374, "num_input_tokens_seen": 311910400, "step": 2380, "train_runtime": 116035.6048, "train_tokens_per_second": 2688.058 }, { "epoch": 1.4538403041825094, "grad_norm": 0.375, "learning_rate": 0.0003680052690763131, "loss": 2.8273, "num_input_tokens_seen": 313221120, "step": 2390, "train_runtime": 116489.8127, "train_tokens_per_second": 2688.828 }, { "epoch": 1.4599239543726235, "grad_norm": 0.349609375, "learning_rate": 0.0003669182398360641, "loss": 2.8182, "num_input_tokens_seen": 314531840, "step": 2400, "train_runtime": 116943.0324, "train_tokens_per_second": 2689.616 }, { "epoch": 1.4660076045627377, "grad_norm": 0.357421875, "learning_rate": 0.00036582837390432326, "loss": 2.8156, "num_input_tokens_seen": 315842560, "step": 2410, "train_runtime": 117395.359, "train_tokens_per_second": 2690.418 }, { "epoch": 1.4720912547528517, "grad_norm": 0.384765625, "learning_rate": 0.00036473569772361514, "loss": 2.852, "num_input_tokens_seen": 317153280, "step": 2420, "train_runtime": 117851.896, "train_tokens_per_second": 2691.117 }, { "epoch": 1.4781749049429658, "grad_norm": 0.361328125, "learning_rate": 0.00036364023780464694, "loss": 2.8335, "num_input_tokens_seen": 318464000, "step": 2430, "train_runtime": 118307.3385, "train_tokens_per_second": 2691.836 }, { "epoch": 1.4842585551330798, "grad_norm": 0.34765625, "learning_rate": 0.0003625420207256656, "loss": 2.8399, "num_input_tokens_seen": 319774720, "step": 2440, "train_runtime": 118760.9215, "train_tokens_per_second": 2692.592 }, { "epoch": 1.4903422053231938, "grad_norm": 0.390625, "learning_rate": 0.00036144107313181224, "loss": 2.8545, "num_input_tokens_seen": 321085440, "step": 2450, "train_runtime": 119212.8391, "train_tokens_per_second": 2693.38 }, { "epoch": 1.496425855513308, "grad_norm": 0.375, "learning_rate": 0.00036033742173447695, "loss": 2.8383, "num_input_tokens_seen": 322396160, "step": 2460, "train_runtime": 119664.4567, "train_tokens_per_second": 2694.168 }, { "epoch": 1.5025095057034221, "grad_norm": 0.375, "learning_rate": 0.0003592310933106495, "loss": 2.8197, "num_input_tokens_seen": 323706880, "step": 2470, "train_runtime": 120117.6232, "train_tokens_per_second": 2694.916 }, { "epoch": 1.5085931558935362, "grad_norm": 0.34765625, "learning_rate": 0.0003581221147022703, "loss": 2.8484, "num_input_tokens_seen": 325017600, "step": 2480, "train_runtime": 120570.0469, "train_tokens_per_second": 2695.674 }, { "epoch": 1.5146768060836502, "grad_norm": 0.359375, "learning_rate": 0.0003570105128155791, "loss": 2.8361, "num_input_tokens_seen": 326328320, "step": 2490, "train_runtime": 121024.5878, "train_tokens_per_second": 2696.38 }, { "epoch": 1.5207604562737642, "grad_norm": 0.357421875, "learning_rate": 0.0003558963146204619, "loss": 2.7998, "num_input_tokens_seen": 327639040, "step": 2500, "train_runtime": 121477.7685, "train_tokens_per_second": 2697.111 }, { "epoch": 1.5268441064638782, "grad_norm": 0.341796875, "learning_rate": 0.0003547795471497971, "loss": 2.8491, "num_input_tokens_seen": 328949760, "step": 2510, "train_runtime": 121932.2878, "train_tokens_per_second": 2697.807 }, { "epoch": 1.5329277566539923, "grad_norm": 0.35546875, "learning_rate": 0.00035366023749879895, "loss": 2.8505, "num_input_tokens_seen": 330260480, "step": 2520, "train_runtime": 122381.2631, "train_tokens_per_second": 2698.62 }, { "epoch": 1.5390114068441063, "grad_norm": 0.34375, "learning_rate": 0.0003525384128243609, "loss": 2.8021, "num_input_tokens_seen": 331571200, "step": 2530, "train_runtime": 122830.6404, "train_tokens_per_second": 2699.418 }, { "epoch": 1.5450950570342206, "grad_norm": 0.32421875, "learning_rate": 0.00035141410034439613, "loss": 2.8371, "num_input_tokens_seen": 332881920, "step": 2540, "train_runtime": 123278.651, "train_tokens_per_second": 2700.24 }, { "epoch": 1.5511787072243346, "grad_norm": 0.341796875, "learning_rate": 0.0003502873273371775, "loss": 2.8455, "num_input_tokens_seen": 334192640, "step": 2550, "train_runtime": 123733.2531, "train_tokens_per_second": 2700.912 }, { "epoch": 1.5572623574144486, "grad_norm": 0.34375, "learning_rate": 0.0003491581211406756, "loss": 2.8163, "num_input_tokens_seen": 335503360, "step": 2560, "train_runtime": 124189.0309, "train_tokens_per_second": 2701.554 }, { "epoch": 1.5633460076045629, "grad_norm": 0.33984375, "learning_rate": 0.00034802650915189554, "loss": 2.8193, "num_input_tokens_seen": 336814080, "step": 2570, "train_runtime": 124645.0232, "train_tokens_per_second": 2702.186 }, { "epoch": 1.569429657794677, "grad_norm": 0.349609375, "learning_rate": 0.0003468925188262121, "loss": 2.8321, "num_input_tokens_seen": 338124800, "step": 2580, "train_runtime": 125098.7972, "train_tokens_per_second": 2702.862 }, { "epoch": 1.575513307984791, "grad_norm": 0.33203125, "learning_rate": 0.00034575617767670377, "loss": 2.8267, "num_input_tokens_seen": 339435520, "step": 2590, "train_runtime": 125553.6767, "train_tokens_per_second": 2703.509 }, { "epoch": 1.581596958174905, "grad_norm": 0.3359375, "learning_rate": 0.0003446175132734852, "loss": 2.8025, "num_input_tokens_seen": 340746240, "step": 2600, "train_runtime": 126010.9594, "train_tokens_per_second": 2704.1 }, { "epoch": 1.587680608365019, "grad_norm": 0.3515625, "learning_rate": 0.0003434765532430382, "loss": 2.7821, "num_input_tokens_seen": 342056960, "step": 2610, "train_runtime": 126467.8994, "train_tokens_per_second": 2704.694 }, { "epoch": 1.593764258555133, "grad_norm": 0.3203125, "learning_rate": 0.00034233332526754137, "loss": 2.8006, "num_input_tokens_seen": 343367680, "step": 2620, "train_runtime": 126924.495, "train_tokens_per_second": 2705.291 }, { "epoch": 1.599847908745247, "grad_norm": 0.322265625, "learning_rate": 0.0003411878570841988, "loss": 2.8201, "num_input_tokens_seen": 344678400, "step": 2630, "train_runtime": 127377.5833, "train_tokens_per_second": 2705.958 }, { "epoch": 1.605931558935361, "grad_norm": 0.369140625, "learning_rate": 0.00034004017648456693, "loss": 2.8014, "num_input_tokens_seen": 345989120, "step": 2640, "train_runtime": 127833.5239, "train_tokens_per_second": 2706.56 }, { "epoch": 1.6120152091254751, "grad_norm": 0.34375, "learning_rate": 0.00033889031131387995, "loss": 2.8407, "num_input_tokens_seen": 347299840, "step": 2650, "train_runtime": 128291.2954, "train_tokens_per_second": 2707.119 }, { "epoch": 1.6180988593155894, "grad_norm": 0.35546875, "learning_rate": 0.0003377382894703748, "loss": 2.8222, "num_input_tokens_seen": 348610560, "step": 2660, "train_runtime": 128747.4205, "train_tokens_per_second": 2707.709 }, { "epoch": 1.6241825095057034, "grad_norm": 0.353515625, "learning_rate": 0.0003365841389046139, "loss": 2.8098, "num_input_tokens_seen": 349921280, "step": 2670, "train_runtime": 129204.1147, "train_tokens_per_second": 2708.283 }, { "epoch": 1.6302661596958175, "grad_norm": 0.38671875, "learning_rate": 0.00033542788761880713, "loss": 2.8376, "num_input_tokens_seen": 351232000, "step": 2680, "train_runtime": 129660.8296, "train_tokens_per_second": 2708.852 }, { "epoch": 1.6363498098859317, "grad_norm": 0.330078125, "learning_rate": 0.00033426956366613254, "loss": 2.8017, "num_input_tokens_seen": 352542720, "step": 2690, "train_runtime": 130118.2978, "train_tokens_per_second": 2709.402 }, { "epoch": 1.6424334600760457, "grad_norm": 0.361328125, "learning_rate": 0.00033310919515005524, "loss": 2.8027, "num_input_tokens_seen": 353853440, "step": 2700, "train_runtime": 130575.9315, "train_tokens_per_second": 2709.944 }, { "epoch": 1.6485171102661598, "grad_norm": 0.353515625, "learning_rate": 0.00033194681022364627, "loss": 2.8082, "num_input_tokens_seen": 355164160, "step": 2710, "train_runtime": 131034.1343, "train_tokens_per_second": 2710.471 }, { "epoch": 1.6546007604562738, "grad_norm": 0.33203125, "learning_rate": 0.00033078243708889904, "loss": 2.7959, "num_input_tokens_seen": 356474880, "step": 2720, "train_runtime": 131492.6924, "train_tokens_per_second": 2710.986 }, { "epoch": 1.6606844106463878, "grad_norm": 0.3671875, "learning_rate": 0.00032961610399604525, "loss": 2.7826, "num_input_tokens_seen": 357785600, "step": 2730, "train_runtime": 131952.3753, "train_tokens_per_second": 2711.475 }, { "epoch": 1.6667680608365019, "grad_norm": 0.328125, "learning_rate": 0.0003284478392428695, "loss": 2.7911, "num_input_tokens_seen": 359096320, "step": 2740, "train_runtime": 132412.9832, "train_tokens_per_second": 2711.942 }, { "epoch": 1.672851711026616, "grad_norm": 0.34375, "learning_rate": 0.0003272776711740224, "loss": 2.8126, "num_input_tokens_seen": 360407040, "step": 2750, "train_runtime": 132871.6286, "train_tokens_per_second": 2712.445 }, { "epoch": 1.67893536121673, "grad_norm": 0.30078125, "learning_rate": 0.00032610562818033374, "loss": 2.7689, "num_input_tokens_seen": 361717760, "step": 2760, "train_runtime": 133332.2259, "train_tokens_per_second": 2712.906 }, { "epoch": 1.685019011406844, "grad_norm": 0.32421875, "learning_rate": 0.0003249317386981224, "loss": 2.773, "num_input_tokens_seen": 363028480, "step": 2770, "train_runtime": 133792.8657, "train_tokens_per_second": 2713.362 }, { "epoch": 1.6911026615969582, "grad_norm": 0.322265625, "learning_rate": 0.00032375603120850764, "loss": 2.7964, "num_input_tokens_seen": 364339200, "step": 2780, "train_runtime": 134252.2551, "train_tokens_per_second": 2713.84 }, { "epoch": 1.6971863117870722, "grad_norm": 0.33203125, "learning_rate": 0.0003225785342367172, "loss": 2.8291, "num_input_tokens_seen": 365649920, "step": 2790, "train_runtime": 134711.3995, "train_tokens_per_second": 2714.321 }, { "epoch": 1.7032699619771863, "grad_norm": 0.3203125, "learning_rate": 0.0003213992763513958, "loss": 2.7957, "num_input_tokens_seen": 366960640, "step": 2800, "train_runtime": 135171.8739, "train_tokens_per_second": 2714.771 }, { "epoch": 1.7093536121673005, "grad_norm": 0.32421875, "learning_rate": 0.00032021828616391146, "loss": 2.8162, "num_input_tokens_seen": 368271360, "step": 2810, "train_runtime": 135631.7281, "train_tokens_per_second": 2715.23 }, { "epoch": 1.7154372623574146, "grad_norm": 0.33203125, "learning_rate": 0.000319035592327662, "loss": 2.7847, "num_input_tokens_seen": 369582080, "step": 2820, "train_runtime": 136091.9477, "train_tokens_per_second": 2715.679 }, { "epoch": 1.7215209125475286, "grad_norm": 0.3515625, "learning_rate": 0.0003178512235373791, "loss": 2.7952, "num_input_tokens_seen": 370892800, "step": 2830, "train_runtime": 136551.491, "train_tokens_per_second": 2716.139 }, { "epoch": 1.7276045627376426, "grad_norm": 0.32421875, "learning_rate": 0.00031666520852843285, "loss": 2.778, "num_input_tokens_seen": 372203520, "step": 2840, "train_runtime": 137010.4585, "train_tokens_per_second": 2716.607 }, { "epoch": 1.7336882129277567, "grad_norm": 0.3203125, "learning_rate": 0.000315477576076134, "loss": 2.7778, "num_input_tokens_seen": 373514240, "step": 2850, "train_runtime": 137473.9898, "train_tokens_per_second": 2716.981 }, { "epoch": 1.7397718631178707, "grad_norm": 0.337890625, "learning_rate": 0.00031428835499503583, "loss": 2.7703, "num_input_tokens_seen": 374824960, "step": 2860, "train_runtime": 137935.0519, "train_tokens_per_second": 2717.402 }, { "epoch": 1.7458555133079847, "grad_norm": 0.326171875, "learning_rate": 0.00031309757413823557, "loss": 2.7657, "num_input_tokens_seen": 376135680, "step": 2870, "train_runtime": 138396.7751, "train_tokens_per_second": 2717.807 }, { "epoch": 1.7519391634980988, "grad_norm": 0.322265625, "learning_rate": 0.0003119052623966738, "loss": 2.7978, "num_input_tokens_seen": 377446400, "step": 2880, "train_runtime": 138858.8271, "train_tokens_per_second": 2718.202 }, { "epoch": 1.7580228136882128, "grad_norm": 0.345703125, "learning_rate": 0.00031071144869843374, "loss": 2.7668, "num_input_tokens_seen": 378757120, "step": 2890, "train_runtime": 139319.9239, "train_tokens_per_second": 2718.614 }, { "epoch": 1.764106463878327, "grad_norm": 0.33203125, "learning_rate": 0.0003095161620080394, "loss": 2.8009, "num_input_tokens_seen": 380067840, "step": 2900, "train_runtime": 139782.9031, "train_tokens_per_second": 2718.987 }, { "epoch": 1.770190114068441, "grad_norm": 0.310546875, "learning_rate": 0.00030831943132575266, "loss": 2.7412, "num_input_tokens_seen": 381378560, "step": 2910, "train_runtime": 140241.6223, "train_tokens_per_second": 2719.439 }, { "epoch": 1.776273764258555, "grad_norm": 0.318359375, "learning_rate": 0.00030712128568687, "loss": 2.7809, "num_input_tokens_seen": 382689280, "step": 2920, "train_runtime": 140700.6087, "train_tokens_per_second": 2719.884 }, { "epoch": 1.7823574144486694, "grad_norm": 0.328125, "learning_rate": 0.00030592175416101767, "loss": 2.7602, "num_input_tokens_seen": 384000000, "step": 2930, "train_runtime": 141160.4567, "train_tokens_per_second": 2720.309 }, { "epoch": 1.7884410646387834, "grad_norm": 0.322265625, "learning_rate": 0.0003047208658514466, "loss": 2.7824, "num_input_tokens_seen": 385310720, "step": 2940, "train_runtime": 141622.4855, "train_tokens_per_second": 2720.689 }, { "epoch": 1.7945247148288974, "grad_norm": 0.328125, "learning_rate": 0.0003035186498943262, "loss": 2.7823, "num_input_tokens_seen": 386621440, "step": 2950, "train_runtime": 142082.5243, "train_tokens_per_second": 2721.105 }, { "epoch": 1.8006083650190114, "grad_norm": 0.326171875, "learning_rate": 0.0003023151354580378, "loss": 2.7379, "num_input_tokens_seen": 387932160, "step": 2960, "train_runtime": 142540.4824, "train_tokens_per_second": 2721.558 }, { "epoch": 1.8066920152091255, "grad_norm": 0.341796875, "learning_rate": 0.00030111035174246615, "loss": 2.7631, "num_input_tokens_seen": 389242880, "step": 2970, "train_runtime": 142998.9456, "train_tokens_per_second": 2721.998 }, { "epoch": 1.8127756653992395, "grad_norm": 0.314453125, "learning_rate": 0.0002999043279782919, "loss": 2.7424, "num_input_tokens_seen": 390553600, "step": 2980, "train_runtime": 143457.6941, "train_tokens_per_second": 2722.43 }, { "epoch": 1.8188593155893535, "grad_norm": 0.328125, "learning_rate": 0.00029869709342628185, "loss": 2.7629, "num_input_tokens_seen": 391864320, "step": 2990, "train_runtime": 143917.5699, "train_tokens_per_second": 2722.839 }, { "epoch": 1.8249429657794676, "grad_norm": 0.337890625, "learning_rate": 0.0002974886773765792, "loss": 2.7772, "num_input_tokens_seen": 393175040, "step": 3000, "train_runtime": 144378.1838, "train_tokens_per_second": 2723.23 }, { "epoch": 1.8310266159695816, "grad_norm": 0.330078125, "learning_rate": 0.00029627910914799266, "loss": 2.7388, "num_input_tokens_seen": 394485760, "step": 3010, "train_runtime": 144840.5, "train_tokens_per_second": 2723.587 }, { "epoch": 1.8371102661596959, "grad_norm": 0.3125, "learning_rate": 0.00029506841808728566, "loss": 2.7696, "num_input_tokens_seen": 395796480, "step": 3020, "train_runtime": 145299.0583, "train_tokens_per_second": 2724.013 }, { "epoch": 1.84319391634981, "grad_norm": 0.3359375, "learning_rate": 0.0002938566335684637, "loss": 2.7519, "num_input_tokens_seen": 397107200, "step": 3030, "train_runtime": 145760.0107, "train_tokens_per_second": 2724.391 }, { "epoch": 1.849277566539924, "grad_norm": 0.3203125, "learning_rate": 0.00029264378499206206, "loss": 2.7506, "num_input_tokens_seen": 398417920, "step": 3040, "train_runtime": 146221.9838, "train_tokens_per_second": 2724.747 }, { "epoch": 1.8553612167300382, "grad_norm": 0.3203125, "learning_rate": 0.0002914299017844324, "loss": 2.7425, "num_input_tokens_seen": 399728640, "step": 3050, "train_runtime": 146684.6189, "train_tokens_per_second": 2725.089 }, { "epoch": 1.8614448669201522, "grad_norm": 0.3203125, "learning_rate": 0.00029021501339702874, "loss": 2.7421, "num_input_tokens_seen": 401039360, "step": 3060, "train_runtime": 147145.5408, "train_tokens_per_second": 2725.461 }, { "epoch": 1.8675285171102662, "grad_norm": 0.3125, "learning_rate": 0.0002889991493056929, "loss": 2.7454, "num_input_tokens_seen": 402350080, "step": 3070, "train_runtime": 147606.196, "train_tokens_per_second": 2725.835 }, { "epoch": 1.8736121673003803, "grad_norm": 0.322265625, "learning_rate": 0.0002877823390099395, "loss": 2.7738, "num_input_tokens_seen": 403660800, "step": 3080, "train_runtime": 148066.6266, "train_tokens_per_second": 2726.211 }, { "epoch": 1.8796958174904943, "grad_norm": 0.330078125, "learning_rate": 0.00028656461203224027, "loss": 2.7429, "num_input_tokens_seen": 404971520, "step": 3090, "train_runtime": 148528.8366, "train_tokens_per_second": 2726.551 }, { "epoch": 1.8857794676806083, "grad_norm": 0.330078125, "learning_rate": 0.0002853459979173074, "loss": 2.7618, "num_input_tokens_seen": 406282240, "step": 3100, "train_runtime": 148990.5945, "train_tokens_per_second": 2726.899 }, { "epoch": 1.8918631178707224, "grad_norm": 0.3359375, "learning_rate": 0.000284126526231377, "loss": 2.7174, "num_input_tokens_seen": 407592960, "step": 3110, "train_runtime": 149437.5489, "train_tokens_per_second": 2727.514 }, { "epoch": 1.8979467680608364, "grad_norm": 0.3203125, "learning_rate": 0.00028290622656149177, "loss": 2.7381, "num_input_tokens_seen": 408903680, "step": 3120, "train_runtime": 149746.6696, "train_tokens_per_second": 2730.636 }, { "epoch": 1.9040304182509504, "grad_norm": 0.302734375, "learning_rate": 0.00028168512851478334, "loss": 2.7346, "num_input_tokens_seen": 410214400, "step": 3130, "train_runtime": 150045.1286, "train_tokens_per_second": 2733.94 }, { "epoch": 1.9101140684410647, "grad_norm": 0.302734375, "learning_rate": 0.00028046326171775326, "loss": 2.7144, "num_input_tokens_seen": 411525120, "step": 3140, "train_runtime": 150338.7507, "train_tokens_per_second": 2737.319 }, { "epoch": 1.9161977186311787, "grad_norm": 0.3203125, "learning_rate": 0.0002792406558155549, "loss": 2.7335, "num_input_tokens_seen": 412835840, "step": 3150, "train_runtime": 150626.8637, "train_tokens_per_second": 2740.785 }, { "epoch": 1.9222813688212927, "grad_norm": 0.318359375, "learning_rate": 0.00027801734047127404, "loss": 2.734, "num_input_tokens_seen": 414146560, "step": 3160, "train_runtime": 150915.2559, "train_tokens_per_second": 2744.233 }, { "epoch": 1.928365019011407, "grad_norm": 0.3203125, "learning_rate": 0.0002767933453652086, "loss": 2.7415, "num_input_tokens_seen": 415457280, "step": 3170, "train_runtime": 151202.2316, "train_tokens_per_second": 2747.693 }, { "epoch": 1.934448669201521, "grad_norm": 0.326171875, "learning_rate": 0.0002755687001941497, "loss": 2.7616, "num_input_tokens_seen": 416768000, "step": 3180, "train_runtime": 151489.837, "train_tokens_per_second": 2751.128 }, { "epoch": 1.940532319391635, "grad_norm": 0.30859375, "learning_rate": 0.0002743434346706599, "loss": 2.721, "num_input_tokens_seen": 418078720, "step": 3190, "train_runtime": 151776.2917, "train_tokens_per_second": 2754.572 }, { "epoch": 1.946615969581749, "grad_norm": 0.31640625, "learning_rate": 0.0002731175785223531, "loss": 2.7402, "num_input_tokens_seen": 419389440, "step": 3200, "train_runtime": 152063.2975, "train_tokens_per_second": 2757.993 }, { "epoch": 1.9526996197718631, "grad_norm": 0.328125, "learning_rate": 0.0002718911614911729, "loss": 2.7048, "num_input_tokens_seen": 420700160, "step": 3210, "train_runtime": 152351.3851, "train_tokens_per_second": 2761.381 }, { "epoch": 1.9587832699619772, "grad_norm": 0.310546875, "learning_rate": 0.00027066421333267135, "loss": 2.7136, "num_input_tokens_seen": 422010880, "step": 3220, "train_runtime": 152637.9334, "train_tokens_per_second": 2764.784 }, { "epoch": 1.9648669201520912, "grad_norm": 0.310546875, "learning_rate": 0.0002694367638152866, "loss": 2.7499, "num_input_tokens_seen": 423321600, "step": 3230, "train_runtime": 152921.9479, "train_tokens_per_second": 2768.22 }, { "epoch": 1.9709505703422052, "grad_norm": 0.31640625, "learning_rate": 0.00026820884271962106, "loss": 2.7487, "num_input_tokens_seen": 424632320, "step": 3240, "train_runtime": 153205.3322, "train_tokens_per_second": 2771.655 }, { "epoch": 1.9770342205323193, "grad_norm": 0.30859375, "learning_rate": 0.00026698047983771844, "loss": 2.7135, "num_input_tokens_seen": 425943040, "step": 3250, "train_runtime": 153491.4373, "train_tokens_per_second": 2775.028 }, { "epoch": 1.9831178707224335, "grad_norm": 0.3125, "learning_rate": 0.00026575170497234135, "loss": 2.7164, "num_input_tokens_seen": 427253760, "step": 3260, "train_runtime": 153778.9801, "train_tokens_per_second": 2778.363 }, { "epoch": 1.9892015209125475, "grad_norm": 0.306640625, "learning_rate": 0.00026452254793624786, "loss": 2.7397, "num_input_tokens_seen": 428564480, "step": 3270, "train_runtime": 154066.1194, "train_tokens_per_second": 2781.692 }, { "epoch": 1.9952851711026616, "grad_norm": 0.322265625, "learning_rate": 0.0002632930385514684, "loss": 2.7232, "num_input_tokens_seen": 429875200, "step": 3280, "train_runtime": 154352.7772, "train_tokens_per_second": 2785.018 }, { "epoch": 2.001216730038023, "grad_norm": 0.53125, "learning_rate": 0.00026206320664858224, "loss": 2.654, "num_input_tokens_seen": 431144960, "step": 3290, "train_runtime": 154631.677, "train_tokens_per_second": 2788.206 }, { "epoch": 2.007300380228137, "grad_norm": 0.462890625, "learning_rate": 0.00026083308206599335, "loss": 2.2631, "num_input_tokens_seen": 432455680, "step": 3300, "train_runtime": 154916.1657, "train_tokens_per_second": 2791.546 }, { "epoch": 2.013384030418251, "grad_norm": 0.431640625, "learning_rate": 0.0002596026946492073, "loss": 2.2231, "num_input_tokens_seen": 433766400, "step": 3310, "train_runtime": 155201.2246, "train_tokens_per_second": 2794.865 }, { "epoch": 2.019467680608365, "grad_norm": 0.44140625, "learning_rate": 0.00025837207425010583, "loss": 2.259, "num_input_tokens_seen": 435077120, "step": 3320, "train_runtime": 155486.5979, "train_tokens_per_second": 2798.165 }, { "epoch": 2.025551330798479, "grad_norm": 0.419921875, "learning_rate": 0.0002571412507262241, "loss": 2.2532, "num_input_tokens_seen": 436387840, "step": 3330, "train_runtime": 155770.6072, "train_tokens_per_second": 2801.477 }, { "epoch": 2.031634980988593, "grad_norm": 0.435546875, "learning_rate": 0.00025591025394002486, "loss": 2.2686, "num_input_tokens_seen": 437698560, "step": 3340, "train_runtime": 156054.2351, "train_tokens_per_second": 2804.785 }, { "epoch": 2.0377186311787074, "grad_norm": 0.40625, "learning_rate": 0.00025467911375817514, "loss": 2.2856, "num_input_tokens_seen": 439009280, "step": 3350, "train_runtime": 156338.1333, "train_tokens_per_second": 2808.075 }, { "epoch": 2.0438022813688215, "grad_norm": 0.39453125, "learning_rate": 0.0002534478600508207, "loss": 2.2971, "num_input_tokens_seen": 440320000, "step": 3360, "train_runtime": 156622.5971, "train_tokens_per_second": 2811.344 }, { "epoch": 2.0498859315589355, "grad_norm": 0.404296875, "learning_rate": 0.00025221652269086177, "loss": 2.251, "num_input_tokens_seen": 441630720, "step": 3370, "train_runtime": 156907.7276, "train_tokens_per_second": 2814.589 }, { "epoch": 2.0559695817490495, "grad_norm": 0.3828125, "learning_rate": 0.0002509851315532283, "loss": 2.2432, "num_input_tokens_seen": 442941440, "step": 3380, "train_runtime": 157192.8368, "train_tokens_per_second": 2817.822 }, { "epoch": 2.0620532319391636, "grad_norm": 0.423828125, "learning_rate": 0.0002497537165141547, "loss": 2.2619, "num_input_tokens_seen": 444252160, "step": 3390, "train_runtime": 157475.922, "train_tokens_per_second": 2821.08 }, { "epoch": 2.0681368821292776, "grad_norm": 0.4453125, "learning_rate": 0.00024852230745045566, "loss": 2.2688, "num_input_tokens_seen": 445562880, "step": 3400, "train_runtime": 157759.8482, "train_tokens_per_second": 2824.311 }, { "epoch": 2.0742205323193916, "grad_norm": 0.396484375, "learning_rate": 0.0002472909342388007, "loss": 2.2595, "num_input_tokens_seen": 446873600, "step": 3410, "train_runtime": 158043.6125, "train_tokens_per_second": 2827.533 }, { "epoch": 2.0803041825095057, "grad_norm": 0.43359375, "learning_rate": 0.0002460596267549894, "loss": 2.2606, "num_input_tokens_seen": 448184320, "step": 3420, "train_runtime": 158327.7009, "train_tokens_per_second": 2830.739 }, { "epoch": 2.0863878326996197, "grad_norm": 0.38671875, "learning_rate": 0.00024482841487322675, "loss": 2.237, "num_input_tokens_seen": 449495040, "step": 3430, "train_runtime": 158615.4509, "train_tokens_per_second": 2833.867 }, { "epoch": 2.0924714828897337, "grad_norm": 0.412109375, "learning_rate": 0.0002435973284653984, "loss": 2.2754, "num_input_tokens_seen": 450805760, "step": 3440, "train_runtime": 158896.2661, "train_tokens_per_second": 2837.107 }, { "epoch": 2.0985551330798478, "grad_norm": 0.42578125, "learning_rate": 0.00024236639740034531, "loss": 2.2738, "num_input_tokens_seen": 452116480, "step": 3450, "train_runtime": 159179.2731, "train_tokens_per_second": 2840.297 }, { "epoch": 2.104638783269962, "grad_norm": 0.44140625, "learning_rate": 0.0002411356515431398, "loss": 2.3094, "num_input_tokens_seen": 453427200, "step": 3460, "train_runtime": 159462.1853, "train_tokens_per_second": 2843.478 }, { "epoch": 2.110722433460076, "grad_norm": 0.4296875, "learning_rate": 0.00023990512075436072, "loss": 2.2732, "num_input_tokens_seen": 454737920, "step": 3470, "train_runtime": 159746.6662, "train_tokens_per_second": 2846.619 }, { "epoch": 2.1168060836501903, "grad_norm": 0.416015625, "learning_rate": 0.00023867483488936847, "loss": 2.2837, "num_input_tokens_seen": 456048640, "step": 3480, "train_runtime": 160029.9994, "train_tokens_per_second": 2849.77 }, { "epoch": 2.1228897338403043, "grad_norm": 0.41796875, "learning_rate": 0.00023744482379758151, "loss": 2.291, "num_input_tokens_seen": 457359360, "step": 3490, "train_runtime": 160313.8968, "train_tokens_per_second": 2852.899 }, { "epoch": 2.1289733840304184, "grad_norm": 0.421875, "learning_rate": 0.00023621511732175156, "loss": 2.2601, "num_input_tokens_seen": 458670080, "step": 3500, "train_runtime": 160597.2641, "train_tokens_per_second": 2856.027 }, { "epoch": 2.1350570342205324, "grad_norm": 0.416015625, "learning_rate": 0.0002349857452972395, "loss": 2.2518, "num_input_tokens_seen": 459980800, "step": 3510, "train_runtime": 160884.5887, "train_tokens_per_second": 2859.073 }, { "epoch": 2.1411406844106464, "grad_norm": 0.412109375, "learning_rate": 0.00023375673755129196, "loss": 2.2733, "num_input_tokens_seen": 461291520, "step": 3520, "train_runtime": 161164.9586, "train_tokens_per_second": 2862.232 }, { "epoch": 2.1472243346007605, "grad_norm": 0.39453125, "learning_rate": 0.00023252812390231725, "loss": 2.2898, "num_input_tokens_seen": 462602240, "step": 3530, "train_runtime": 161447.4137, "train_tokens_per_second": 2865.343 }, { "epoch": 2.1533079847908745, "grad_norm": 0.404296875, "learning_rate": 0.000231299934159162, "loss": 2.3022, "num_input_tokens_seen": 463912960, "step": 3540, "train_runtime": 161731.5367, "train_tokens_per_second": 2868.414 }, { "epoch": 2.1593916349809885, "grad_norm": 0.419921875, "learning_rate": 0.000230072198120388, "loss": 2.2692, "num_input_tokens_seen": 465223680, "step": 3550, "train_runtime": 162015.8766, "train_tokens_per_second": 2871.47 }, { "epoch": 2.1654752851711025, "grad_norm": 0.40625, "learning_rate": 0.00022884494557354938, "loss": 2.2573, "num_input_tokens_seen": 466534400, "step": 3560, "train_runtime": 162300.8248, "train_tokens_per_second": 2874.504 }, { "epoch": 2.1715589353612166, "grad_norm": 0.40625, "learning_rate": 0.00022761820629446915, "loss": 2.3002, "num_input_tokens_seen": 467845120, "step": 3570, "train_runtime": 162584.845, "train_tokens_per_second": 2877.544 }, { "epoch": 2.1776425855513306, "grad_norm": 0.412109375, "learning_rate": 0.00022639201004651796, "loss": 2.2796, "num_input_tokens_seen": 469155840, "step": 3580, "train_runtime": 162865.8342, "train_tokens_per_second": 2880.628 }, { "epoch": 2.183726235741445, "grad_norm": 0.41015625, "learning_rate": 0.00022516638657989104, "loss": 2.2629, "num_input_tokens_seen": 470466560, "step": 3590, "train_runtime": 163150.1327, "train_tokens_per_second": 2883.642 }, { "epoch": 2.189809885931559, "grad_norm": 0.40234375, "learning_rate": 0.00022394136563088664, "loss": 2.2593, "num_input_tokens_seen": 471777280, "step": 3600, "train_runtime": 163433.4214, "train_tokens_per_second": 2886.663 }, { "epoch": 2.195893536121673, "grad_norm": 0.44140625, "learning_rate": 0.00022271697692118474, "loss": 2.2636, "num_input_tokens_seen": 473088000, "step": 3610, "train_runtime": 163714.21, "train_tokens_per_second": 2889.719 }, { "epoch": 2.201977186311787, "grad_norm": 0.390625, "learning_rate": 0.00022149325015712572, "loss": 2.3206, "num_input_tokens_seen": 474398720, "step": 3620, "train_runtime": 163996.6786, "train_tokens_per_second": 2892.734 }, { "epoch": 2.208060836501901, "grad_norm": 0.390625, "learning_rate": 0.0002202702150289898, "loss": 2.2987, "num_input_tokens_seen": 475709440, "step": 3630, "train_runtime": 164277.0917, "train_tokens_per_second": 2895.775 }, { "epoch": 2.2141444866920152, "grad_norm": 0.390625, "learning_rate": 0.0002190479012102766, "loss": 2.304, "num_input_tokens_seen": 477020160, "step": 3640, "train_runtime": 164560.103, "train_tokens_per_second": 2898.759 }, { "epoch": 2.2202281368821293, "grad_norm": 0.40234375, "learning_rate": 0.00021782633835698513, "loss": 2.303, "num_input_tokens_seen": 478330880, "step": 3650, "train_runtime": 164843.6011, "train_tokens_per_second": 2901.725 }, { "epoch": 2.2263117870722433, "grad_norm": 0.40625, "learning_rate": 0.00021660555610689432, "loss": 2.297, "num_input_tokens_seen": 479641600, "step": 3660, "train_runtime": 165127.0401, "train_tokens_per_second": 2904.682 }, { "epoch": 2.2323954372623573, "grad_norm": 0.380859375, "learning_rate": 0.00021538558407884428, "loss": 2.3029, "num_input_tokens_seen": 480952320, "step": 3670, "train_runtime": 165410.064, "train_tokens_per_second": 2907.636 }, { "epoch": 2.2384790874524714, "grad_norm": 0.39453125, "learning_rate": 0.0002141664518720168, "loss": 2.2902, "num_input_tokens_seen": 482263040, "step": 3680, "train_runtime": 165690.9192, "train_tokens_per_second": 2910.618 }, { "epoch": 2.2445627376425854, "grad_norm": 0.396484375, "learning_rate": 0.00021294818906521826, "loss": 2.2598, "num_input_tokens_seen": 483573760, "step": 3690, "train_runtime": 165969.7636, "train_tokens_per_second": 2913.626 }, { "epoch": 2.2506463878326994, "grad_norm": 0.392578125, "learning_rate": 0.0002117308252161614, "loss": 2.2521, "num_input_tokens_seen": 484884480, "step": 3700, "train_runtime": 166249.1147, "train_tokens_per_second": 2916.614 }, { "epoch": 2.2567300380228135, "grad_norm": 0.392578125, "learning_rate": 0.000210514389860748, "loss": 2.298, "num_input_tokens_seen": 486195200, "step": 3710, "train_runtime": 166531.0263, "train_tokens_per_second": 2919.547 }, { "epoch": 2.262813688212928, "grad_norm": 0.3828125, "learning_rate": 0.00020929891251235288, "loss": 2.2704, "num_input_tokens_seen": 487505920, "step": 3720, "train_runtime": 166816.4218, "train_tokens_per_second": 2922.41 }, { "epoch": 2.268897338403042, "grad_norm": 0.380859375, "learning_rate": 0.00020808442266110754, "loss": 2.2685, "num_input_tokens_seen": 488816640, "step": 3730, "train_runtime": 167139.2928, "train_tokens_per_second": 2924.606 }, { "epoch": 2.274980988593156, "grad_norm": 0.38671875, "learning_rate": 0.00020687094977318426, "loss": 2.2528, "num_input_tokens_seen": 490127360, "step": 3740, "train_runtime": 167541.2573, "train_tokens_per_second": 2925.413 }, { "epoch": 2.28106463878327, "grad_norm": 0.40234375, "learning_rate": 0.00020565852329008198, "loss": 2.2887, "num_input_tokens_seen": 491438080, "step": 3750, "train_runtime": 167945.8138, "train_tokens_per_second": 2926.17 }, { "epoch": 2.287148288973384, "grad_norm": 0.376953125, "learning_rate": 0.00020444717262791155, "loss": 2.2814, "num_input_tokens_seen": 492748800, "step": 3760, "train_runtime": 168342.5277, "train_tokens_per_second": 2927.061 }, { "epoch": 2.293231939163498, "grad_norm": 0.388671875, "learning_rate": 0.00020323692717668163, "loss": 2.2865, "num_input_tokens_seen": 494059520, "step": 3770, "train_runtime": 168747.5236, "train_tokens_per_second": 2927.803 }, { "epoch": 2.299315589353612, "grad_norm": 0.376953125, "learning_rate": 0.00020202781629958654, "loss": 2.286, "num_input_tokens_seen": 495370240, "step": 3780, "train_runtime": 169145.8982, "train_tokens_per_second": 2928.657 }, { "epoch": 2.305399239543726, "grad_norm": 0.3984375, "learning_rate": 0.0002008198693322933, "loss": 2.2767, "num_input_tokens_seen": 496680960, "step": 3790, "train_runtime": 169485.8374, "train_tokens_per_second": 2930.516 }, { "epoch": 2.31148288973384, "grad_norm": 0.392578125, "learning_rate": 0.00019961311558222957, "loss": 2.2803, "num_input_tokens_seen": 497991680, "step": 3800, "train_runtime": 169783.4763, "train_tokens_per_second": 2933.099 }, { "epoch": 2.3175665399239542, "grad_norm": 0.392578125, "learning_rate": 0.00019840758432787325, "loss": 2.2858, "num_input_tokens_seen": 499302400, "step": 3810, "train_runtime": 170120.6901, "train_tokens_per_second": 2934.989 }, { "epoch": 2.3236501901140683, "grad_norm": 0.41796875, "learning_rate": 0.00019720330481804194, "loss": 2.2956, "num_input_tokens_seen": 500613120, "step": 3820, "train_runtime": 170509.9773, "train_tokens_per_second": 2935.976 }, { "epoch": 2.3297338403041827, "grad_norm": 0.408203125, "learning_rate": 0.00019600030627118286, "loss": 2.2794, "num_input_tokens_seen": 501923840, "step": 3830, "train_runtime": 170909.5246, "train_tokens_per_second": 2936.781 }, { "epoch": 2.3358174904942968, "grad_norm": 0.380859375, "learning_rate": 0.00019479861787466456, "loss": 2.2827, "num_input_tokens_seen": 503234560, "step": 3840, "train_runtime": 171311.8975, "train_tokens_per_second": 2937.534 }, { "epoch": 2.341901140684411, "grad_norm": 0.390625, "learning_rate": 0.00019359826878406823, "loss": 2.2841, "num_input_tokens_seen": 504545280, "step": 3850, "train_runtime": 171715.6137, "train_tokens_per_second": 2938.261 }, { "epoch": 2.347984790874525, "grad_norm": 0.375, "learning_rate": 0.00019239928812248073, "loss": 2.2691, "num_input_tokens_seen": 505856000, "step": 3860, "train_runtime": 172119.6283, "train_tokens_per_second": 2938.979 }, { "epoch": 2.354068441064639, "grad_norm": 0.390625, "learning_rate": 0.00019120170497978806, "loss": 2.2681, "num_input_tokens_seen": 507166720, "step": 3870, "train_runtime": 172448.2534, "train_tokens_per_second": 2940.979 }, { "epoch": 2.360152091254753, "grad_norm": 0.404296875, "learning_rate": 0.0001900055484119688, "loss": 2.2728, "num_input_tokens_seen": 508477440, "step": 3880, "train_runtime": 172754.8765, "train_tokens_per_second": 2943.346 }, { "epoch": 2.366235741444867, "grad_norm": 0.365234375, "learning_rate": 0.0001888108474403903, "loss": 2.2809, "num_input_tokens_seen": 509788160, "step": 3890, "train_runtime": 173072.7558, "train_tokens_per_second": 2945.514 }, { "epoch": 2.372319391634981, "grad_norm": 0.396484375, "learning_rate": 0.00018761763105110368, "loss": 2.2759, "num_input_tokens_seen": 511098880, "step": 3900, "train_runtime": 173461.4391, "train_tokens_per_second": 2946.47 }, { "epoch": 2.378403041825095, "grad_norm": 0.38671875, "learning_rate": 0.0001864259281941408, "loss": 2.2714, "num_input_tokens_seen": 512409600, "step": 3910, "train_runtime": 173862.6385, "train_tokens_per_second": 2947.209 }, { "epoch": 2.384486692015209, "grad_norm": 0.390625, "learning_rate": 0.00018523576778281203, "loss": 2.3315, "num_input_tokens_seen": 513720320, "step": 3920, "train_runtime": 174260.0105, "train_tokens_per_second": 2948.01 }, { "epoch": 2.390570342205323, "grad_norm": 0.365234375, "learning_rate": 0.00018404717869300474, "loss": 2.2851, "num_input_tokens_seen": 515031040, "step": 3930, "train_runtime": 174662.9573, "train_tokens_per_second": 2948.714 }, { "epoch": 2.396653992395437, "grad_norm": 0.400390625, "learning_rate": 0.00018286018976248241, "loss": 2.2702, "num_input_tokens_seen": 516341760, "step": 3940, "train_runtime": 175072.4087, "train_tokens_per_second": 2949.304 }, { "epoch": 2.402737642585551, "grad_norm": 0.3671875, "learning_rate": 0.00018167482979018524, "loss": 2.2418, "num_input_tokens_seen": 517652480, "step": 3950, "train_runtime": 175479.8988, "train_tokens_per_second": 2949.925 }, { "epoch": 2.4088212927756656, "grad_norm": 0.400390625, "learning_rate": 0.00018049112753553155, "loss": 2.2468, "num_input_tokens_seen": 518963200, "step": 3960, "train_runtime": 175887.4195, "train_tokens_per_second": 2950.542 }, { "epoch": 2.4149049429657796, "grad_norm": 0.40625, "learning_rate": 0.00017930911171771936, "loss": 2.2478, "num_input_tokens_seen": 520273920, "step": 3970, "train_runtime": 176291.8991, "train_tokens_per_second": 2951.207 }, { "epoch": 2.4209885931558937, "grad_norm": 0.37890625, "learning_rate": 0.00017812881101503055, "loss": 2.2586, "num_input_tokens_seen": 521584640, "step": 3980, "train_runtime": 176698.2006, "train_tokens_per_second": 2951.839 }, { "epoch": 2.4270722433460077, "grad_norm": 0.400390625, "learning_rate": 0.00017695025406413433, "loss": 2.2691, "num_input_tokens_seen": 522895360, "step": 3990, "train_runtime": 177101.7892, "train_tokens_per_second": 2952.513 }, { "epoch": 2.4331558935361217, "grad_norm": 0.388671875, "learning_rate": 0.00017577346945939266, "loss": 2.2809, "num_input_tokens_seen": 524206080, "step": 4000, "train_runtime": 177508.9278, "train_tokens_per_second": 2953.125 }, { "epoch": 2.4392395437262357, "grad_norm": 0.37890625, "learning_rate": 0.00017459848575216653, "loss": 2.26, "num_input_tokens_seen": 525516800, "step": 4010, "train_runtime": 177913.3706, "train_tokens_per_second": 2953.779 }, { "epoch": 2.44532319391635, "grad_norm": 0.369140625, "learning_rate": 0.00017342533145012334, "loss": 2.2354, "num_input_tokens_seen": 526827520, "step": 4020, "train_runtime": 178319.7473, "train_tokens_per_second": 2954.398 }, { "epoch": 2.451406844106464, "grad_norm": 0.3828125, "learning_rate": 0.00017225403501654497, "loss": 2.2772, "num_input_tokens_seen": 528138240, "step": 4030, "train_runtime": 178695.0138, "train_tokens_per_second": 2955.529 }, { "epoch": 2.457490494296578, "grad_norm": 0.390625, "learning_rate": 0.00017108462486963743, "loss": 2.2981, "num_input_tokens_seen": 529448960, "step": 4040, "train_runtime": 179008.7387, "train_tokens_per_second": 2957.671 }, { "epoch": 2.463574144486692, "grad_norm": 0.3828125, "learning_rate": 0.00016991712938184154, "loss": 2.2367, "num_input_tokens_seen": 530759680, "step": 4050, "train_runtime": 179328.5813, "train_tokens_per_second": 2959.705 }, { "epoch": 2.469657794676806, "grad_norm": 0.380859375, "learning_rate": 0.00016875157687914374, "loss": 2.2536, "num_input_tokens_seen": 532070400, "step": 4060, "train_runtime": 179694.1448, "train_tokens_per_second": 2960.978 }, { "epoch": 2.4757414448669204, "grad_norm": 0.400390625, "learning_rate": 0.00016758799564039011, "loss": 2.253, "num_input_tokens_seen": 533381120, "step": 4070, "train_runtime": 180006.4269, "train_tokens_per_second": 2963.123 }, { "epoch": 2.4818250950570344, "grad_norm": 0.38671875, "learning_rate": 0.00016642641389659891, "loss": 2.2397, "num_input_tokens_seen": 534691840, "step": 4080, "train_runtime": 180374.7182, "train_tokens_per_second": 2964.339 }, { "epoch": 2.4879087452471484, "grad_norm": 0.37890625, "learning_rate": 0.00016526685983027676, "loss": 2.2834, "num_input_tokens_seen": 536002560, "step": 4090, "train_runtime": 180763.2045, "train_tokens_per_second": 2965.219 }, { "epoch": 2.4939923954372625, "grad_norm": 0.388671875, "learning_rate": 0.00016410936157473428, "loss": 2.2324, "num_input_tokens_seen": 537313280, "step": 4100, "train_runtime": 181162.6654, "train_tokens_per_second": 2965.916 }, { "epoch": 2.5000760456273765, "grad_norm": 0.369140625, "learning_rate": 0.00016295394721340352, "loss": 2.2596, "num_input_tokens_seen": 538624000, "step": 4110, "train_runtime": 181549.9797, "train_tokens_per_second": 2966.808 }, { "epoch": 2.5061596958174905, "grad_norm": 0.390625, "learning_rate": 0.0001618006447791568, "loss": 2.2531, "num_input_tokens_seen": 539934720, "step": 4120, "train_runtime": 181949.9324, "train_tokens_per_second": 2967.491 }, { "epoch": 2.5122433460076046, "grad_norm": 0.365234375, "learning_rate": 0.00016064948225362653, "loss": 2.2416, "num_input_tokens_seen": 541245440, "step": 4130, "train_runtime": 182347.8217, "train_tokens_per_second": 2968.203 }, { "epoch": 2.5183269961977186, "grad_norm": 0.37890625, "learning_rate": 0.00015950048756652613, "loss": 2.2866, "num_input_tokens_seen": 542556160, "step": 4140, "train_runtime": 182739.9129, "train_tokens_per_second": 2969.007 }, { "epoch": 2.5244106463878326, "grad_norm": 0.376953125, "learning_rate": 0.00015835368859497262, "loss": 2.2565, "num_input_tokens_seen": 543866880, "step": 4150, "train_runtime": 183056.8824, "train_tokens_per_second": 2971.027 }, { "epoch": 2.5304942965779467, "grad_norm": 0.388671875, "learning_rate": 0.00015720911316281035, "loss": 2.278, "num_input_tokens_seen": 545177600, "step": 4160, "train_runtime": 183352.523, "train_tokens_per_second": 2973.385 }, { "epoch": 2.5365779467680607, "grad_norm": 0.392578125, "learning_rate": 0.00015606678903993526, "loss": 2.2587, "num_input_tokens_seen": 546488320, "step": 4170, "train_runtime": 183645.3012, "train_tokens_per_second": 2975.782 }, { "epoch": 2.542661596958175, "grad_norm": 0.38671875, "learning_rate": 0.00015492674394162212, "loss": 2.2541, "num_input_tokens_seen": 547799040, "step": 4180, "train_runtime": 183930.1192, "train_tokens_per_second": 2978.3 }, { "epoch": 2.5487452471482888, "grad_norm": 0.365234375, "learning_rate": 0.00015378900552785147, "loss": 2.2852, "num_input_tokens_seen": 549109760, "step": 4190, "train_runtime": 184214.7892, "train_tokens_per_second": 2980.813 }, { "epoch": 2.5548288973384032, "grad_norm": 0.41796875, "learning_rate": 0.0001526536014026384, "loss": 2.2351, "num_input_tokens_seen": 550420480, "step": 4200, "train_runtime": 184498.9282, "train_tokens_per_second": 2983.326 }, { "epoch": 2.5609125475285173, "grad_norm": 0.373046875, "learning_rate": 0.00015152055911336331, "loss": 2.2378, "num_input_tokens_seen": 551731200, "step": 4210, "train_runtime": 184784.9113, "train_tokens_per_second": 2985.802 }, { "epoch": 2.5669961977186313, "grad_norm": 0.37890625, "learning_rate": 0.0001503899061501034, "loss": 2.2727, "num_input_tokens_seen": 553041920, "step": 4220, "train_runtime": 185071.4629, "train_tokens_per_second": 2988.261 }, { "epoch": 2.5730798479087453, "grad_norm": 0.369140625, "learning_rate": 0.00014926166994496526, "loss": 2.2601, "num_input_tokens_seen": 554352640, "step": 4230, "train_runtime": 185356.131, "train_tokens_per_second": 2990.743 }, { "epoch": 2.5791634980988594, "grad_norm": 0.388671875, "learning_rate": 0.00014813587787141985, "loss": 2.2717, "num_input_tokens_seen": 555663360, "step": 4240, "train_runtime": 185642.9747, "train_tokens_per_second": 2993.183 }, { "epoch": 2.5852471482889734, "grad_norm": 0.40625, "learning_rate": 0.00014701255724363846, "loss": 2.2139, "num_input_tokens_seen": 556974080, "step": 4250, "train_runtime": 185930.8672, "train_tokens_per_second": 2995.598 }, { "epoch": 2.5913307984790874, "grad_norm": 0.384765625, "learning_rate": 0.0001458917353158291, "loss": 2.2556, "num_input_tokens_seen": 558284800, "step": 4260, "train_runtime": 186218.2786, "train_tokens_per_second": 2998.013 }, { "epoch": 2.5974144486692015, "grad_norm": 0.400390625, "learning_rate": 0.00014477343928157637, "loss": 2.2514, "num_input_tokens_seen": 559595520, "step": 4270, "train_runtime": 186506.0602, "train_tokens_per_second": 3000.415 }, { "epoch": 2.6034980988593155, "grad_norm": 0.408203125, "learning_rate": 0.00014365769627318095, "loss": 2.2772, "num_input_tokens_seen": 560906240, "step": 4280, "train_runtime": 186795.2444, "train_tokens_per_second": 3002.787 }, { "epoch": 2.6095817490494295, "grad_norm": 0.380859375, "learning_rate": 0.00014254453336100164, "loss": 2.2335, "num_input_tokens_seen": 562216960, "step": 4290, "train_runtime": 187082.6925, "train_tokens_per_second": 3005.179 }, { "epoch": 2.6156653992395436, "grad_norm": 0.388671875, "learning_rate": 0.00014143397755279853, "loss": 2.2124, "num_input_tokens_seen": 563527680, "step": 4300, "train_runtime": 187370.1439, "train_tokens_per_second": 3007.564 }, { "epoch": 2.621749049429658, "grad_norm": 0.388671875, "learning_rate": 0.00014032605579307746, "loss": 2.242, "num_input_tokens_seen": 564838400, "step": 4310, "train_runtime": 187656.4196, "train_tokens_per_second": 3009.96 }, { "epoch": 2.6278326996197716, "grad_norm": 0.3828125, "learning_rate": 0.00013922079496243652, "loss": 2.2601, "num_input_tokens_seen": 566149120, "step": 4320, "train_runtime": 187943.3185, "train_tokens_per_second": 3012.34 }, { "epoch": 2.633916349809886, "grad_norm": 0.38671875, "learning_rate": 0.00013811822187691442, "loss": 2.2565, "num_input_tokens_seen": 567459840, "step": 4330, "train_runtime": 188229.6957, "train_tokens_per_second": 3014.72 }, { "epoch": 2.64, "grad_norm": 0.380859375, "learning_rate": 0.00013701836328733848, "loss": 2.2579, "num_input_tokens_seen": 568770560, "step": 4340, "train_runtime": 188511.174, "train_tokens_per_second": 3017.172 }, { "epoch": 2.646083650190114, "grad_norm": 0.376953125, "learning_rate": 0.00013592124587867699, "loss": 2.2485, "num_input_tokens_seen": 570081280, "step": 4350, "train_runtime": 188795.3102, "train_tokens_per_second": 3019.573 }, { "epoch": 2.652167300380228, "grad_norm": 0.380859375, "learning_rate": 0.00013482689626939098, "loss": 2.2866, "num_input_tokens_seen": 571392000, "step": 4360, "train_runtime": 189080.8835, "train_tokens_per_second": 3021.945 }, { "epoch": 2.658250950570342, "grad_norm": 0.373046875, "learning_rate": 0.00013373534101078879, "loss": 2.2841, "num_input_tokens_seen": 572702720, "step": 4370, "train_runtime": 189363.4231, "train_tokens_per_second": 3024.358 }, { "epoch": 2.6643346007604563, "grad_norm": 0.365234375, "learning_rate": 0.00013264660658638156, "loss": 2.2631, "num_input_tokens_seen": 574013440, "step": 4380, "train_runtime": 189647.906, "train_tokens_per_second": 3026.732 }, { "epoch": 2.6704182509505703, "grad_norm": 0.3828125, "learning_rate": 0.00013156071941124097, "loss": 2.2472, "num_input_tokens_seen": 575324160, "step": 4390, "train_runtime": 189933.2699, "train_tokens_per_second": 3029.086 }, { "epoch": 2.6765019011406843, "grad_norm": 0.36328125, "learning_rate": 0.00013047770583135806, "loss": 2.2455, "num_input_tokens_seen": 576634880, "step": 4400, "train_runtime": 190217.0142, "train_tokens_per_second": 3031.458 }, { "epoch": 2.6825855513307983, "grad_norm": 0.376953125, "learning_rate": 0.00012939759212300423, "loss": 2.2443, "num_input_tokens_seen": 577945600, "step": 4410, "train_runtime": 190502.7513, "train_tokens_per_second": 3033.791 }, { "epoch": 2.688669201520913, "grad_norm": 0.380859375, "learning_rate": 0.00012832040449209397, "loss": 2.235, "num_input_tokens_seen": 579256320, "step": 4420, "train_runtime": 190784.7777, "train_tokens_per_second": 3036.177 }, { "epoch": 2.6947528517110264, "grad_norm": 0.373046875, "learning_rate": 0.00012724616907354831, "loss": 2.254, "num_input_tokens_seen": 580567040, "step": 4430, "train_runtime": 191069.3239, "train_tokens_per_second": 3038.515 }, { "epoch": 2.700836501901141, "grad_norm": 0.375, "learning_rate": 0.00012617491193066138, "loss": 2.234, "num_input_tokens_seen": 581877760, "step": 4440, "train_runtime": 191351.4149, "train_tokens_per_second": 3040.886 }, { "epoch": 2.706920152091255, "grad_norm": 0.412109375, "learning_rate": 0.00012510665905446786, "loss": 2.2605, "num_input_tokens_seen": 583188480, "step": 4450, "train_runtime": 191635.3856, "train_tokens_per_second": 3043.219 }, { "epoch": 2.713003802281369, "grad_norm": 0.435546875, "learning_rate": 0.00012404143636311234, "loss": 2.2754, "num_input_tokens_seen": 584499200, "step": 4460, "train_runtime": 191919.5565, "train_tokens_per_second": 3045.543 }, { "epoch": 2.719087452471483, "grad_norm": 0.384765625, "learning_rate": 0.0001229792697012205, "loss": 2.2325, "num_input_tokens_seen": 585809920, "step": 4470, "train_runtime": 192203.536, "train_tokens_per_second": 3047.862 }, { "epoch": 2.725171102661597, "grad_norm": 0.435546875, "learning_rate": 0.0001219201848392722, "loss": 2.2679, "num_input_tokens_seen": 587120640, "step": 4480, "train_runtime": 192486.4046, "train_tokens_per_second": 3050.193 }, { "epoch": 2.731254752851711, "grad_norm": 0.462890625, "learning_rate": 0.0001208642074729758, "loss": 2.2326, "num_input_tokens_seen": 588431360, "step": 4490, "train_runtime": 192772.2556, "train_tokens_per_second": 3052.469 }, { "epoch": 2.737338403041825, "grad_norm": 0.369140625, "learning_rate": 0.00011981136322264527, "loss": 2.2257, "num_input_tokens_seen": 589742080, "step": 4500, "train_runtime": 193056.7552, "train_tokens_per_second": 3054.76 }, { "epoch": 2.743422053231939, "grad_norm": 0.392578125, "learning_rate": 0.00011876167763257859, "loss": 2.2207, "num_input_tokens_seen": 591052800, "step": 4510, "train_runtime": 193339.6912, "train_tokens_per_second": 3057.069 }, { "epoch": 2.749505703422053, "grad_norm": 0.384765625, "learning_rate": 0.00011771517617043723, "loss": 2.2161, "num_input_tokens_seen": 592363520, "step": 4520, "train_runtime": 193619.5072, "train_tokens_per_second": 3059.421 }, { "epoch": 2.755589353612167, "grad_norm": 0.3828125, "learning_rate": 0.0001166718842266292, "loss": 2.2141, "num_input_tokens_seen": 593674240, "step": 4530, "train_runtime": 193900.0918, "train_tokens_per_second": 3061.753 }, { "epoch": 2.761673003802281, "grad_norm": 0.390625, "learning_rate": 0.00011563182711369216, "loss": 2.2488, "num_input_tokens_seen": 594984960, "step": 4540, "train_runtime": 194181.2319, "train_tokens_per_second": 3064.07 }, { "epoch": 2.7677566539923957, "grad_norm": 0.404296875, "learning_rate": 0.00011459503006568026, "loss": 2.236, "num_input_tokens_seen": 596295680, "step": 4550, "train_runtime": 194461.9664, "train_tokens_per_second": 3066.387 }, { "epoch": 2.7738403041825093, "grad_norm": 0.416015625, "learning_rate": 0.00011356151823755109, "loss": 2.2273, "num_input_tokens_seen": 597606400, "step": 4560, "train_runtime": 194741.6653, "train_tokens_per_second": 3068.714 }, { "epoch": 2.7799239543726237, "grad_norm": 0.419921875, "learning_rate": 0.00011253131670455547, "loss": 2.2011, "num_input_tokens_seen": 598917120, "step": 4570, "train_runtime": 195021.7225, "train_tokens_per_second": 3071.028 }, { "epoch": 2.7860076045627378, "grad_norm": 0.43359375, "learning_rate": 0.00011150445046162949, "loss": 2.2139, "num_input_tokens_seen": 600227840, "step": 4580, "train_runtime": 195305.1746, "train_tokens_per_second": 3073.282 }, { "epoch": 2.792091254752852, "grad_norm": 0.388671875, "learning_rate": 0.00011048094442278775, "loss": 2.2377, "num_input_tokens_seen": 601538560, "step": 4590, "train_runtime": 195586.2309, "train_tokens_per_second": 3075.567 }, { "epoch": 2.798174904942966, "grad_norm": 0.42578125, "learning_rate": 0.000109460823420519, "loss": 2.2328, "num_input_tokens_seen": 602849280, "step": 4600, "train_runtime": 195869.0835, "train_tokens_per_second": 3077.817 }, { "epoch": 2.80425855513308, "grad_norm": 0.462890625, "learning_rate": 0.00010844411220518357, "loss": 2.2193, "num_input_tokens_seen": 604160000, "step": 4610, "train_runtime": 196149.6476, "train_tokens_per_second": 3080.097 }, { "epoch": 2.810342205323194, "grad_norm": 0.400390625, "learning_rate": 0.00010743083544441304, "loss": 2.2007, "num_input_tokens_seen": 605470720, "step": 4620, "train_runtime": 196431.3842, "train_tokens_per_second": 3082.352 }, { "epoch": 2.816425855513308, "grad_norm": 0.375, "learning_rate": 0.00010642101772251126, "loss": 2.2269, "num_input_tokens_seen": 606781440, "step": 4630, "train_runtime": 196713.1962, "train_tokens_per_second": 3084.6 }, { "epoch": 2.822509505703422, "grad_norm": 0.3828125, "learning_rate": 0.00010541468353985878, "loss": 2.2308, "num_input_tokens_seen": 608092160, "step": 4640, "train_runtime": 196995.2329, "train_tokens_per_second": 3086.837 }, { "epoch": 2.828593155893536, "grad_norm": 0.421875, "learning_rate": 0.0001044118573123177, "loss": 2.2212, "num_input_tokens_seen": 609402880, "step": 4650, "train_runtime": 197277.9382, "train_tokens_per_second": 3089.057 }, { "epoch": 2.83467680608365, "grad_norm": 0.3828125, "learning_rate": 0.00010341256337063917, "loss": 2.251, "num_input_tokens_seen": 610713600, "step": 4660, "train_runtime": 197559.8157, "train_tokens_per_second": 3091.285 }, { "epoch": 2.840760456273764, "grad_norm": 0.3984375, "learning_rate": 0.00010241682595987378, "loss": 2.232, "num_input_tokens_seen": 612024320, "step": 4670, "train_runtime": 197841.9689, "train_tokens_per_second": 3093.501 }, { "epoch": 2.8468441064638785, "grad_norm": 0.369140625, "learning_rate": 0.00010142466923878274, "loss": 2.2362, "num_input_tokens_seen": 613335040, "step": 4680, "train_runtime": 198123.472, "train_tokens_per_second": 3095.721 }, { "epoch": 2.852927756653992, "grad_norm": 0.380859375, "learning_rate": 0.00010043611727925194, "loss": 2.2195, "num_input_tokens_seen": 614645760, "step": 4690, "train_runtime": 198406.0097, "train_tokens_per_second": 3097.919 }, { "epoch": 2.8590114068441066, "grad_norm": 0.380859375, "learning_rate": 9.945119406570793e-05, "loss": 2.2289, "num_input_tokens_seen": 615956480, "step": 4700, "train_runtime": 198686.4134, "train_tokens_per_second": 3100.144 }, { "epoch": 2.8650950570342206, "grad_norm": 0.388671875, "learning_rate": 9.846992349453602e-05, "loss": 2.2286, "num_input_tokens_seen": 617267200, "step": 4710, "train_runtime": 198964.9676, "train_tokens_per_second": 3102.391 }, { "epoch": 2.8711787072243347, "grad_norm": 0.376953125, "learning_rate": 9.749232937350011e-05, "loss": 2.2129, "num_input_tokens_seen": 618577920, "step": 4720, "train_runtime": 199243.6329, "train_tokens_per_second": 3104.631 }, { "epoch": 2.8772623574144487, "grad_norm": 0.369140625, "learning_rate": 9.651843542116592e-05, "loss": 2.206, "num_input_tokens_seen": 619888640, "step": 4730, "train_runtime": 199525.2636, "train_tokens_per_second": 3106.818 }, { "epoch": 2.8833460076045627, "grad_norm": 0.388671875, "learning_rate": 9.554826526632476e-05, "loss": 2.1995, "num_input_tokens_seen": 621199360, "step": 4740, "train_runtime": 199805.6604, "train_tokens_per_second": 3109.018 }, { "epoch": 2.8894296577946768, "grad_norm": 0.37890625, "learning_rate": 9.458184244742033e-05, "loss": 2.231, "num_input_tokens_seen": 622510080, "step": 4750, "train_runtime": 200085.6118, "train_tokens_per_second": 3111.219 }, { "epoch": 2.895513307984791, "grad_norm": 0.384765625, "learning_rate": 9.3619190411978e-05, "loss": 2.2128, "num_input_tokens_seen": 623820800, "step": 4760, "train_runtime": 200366.2522, "train_tokens_per_second": 3113.403 }, { "epoch": 2.901596958174905, "grad_norm": 0.41015625, "learning_rate": 9.266033251603564e-05, "loss": 2.2127, "num_input_tokens_seen": 625131520, "step": 4770, "train_runtime": 200645.7105, "train_tokens_per_second": 3115.599 }, { "epoch": 2.907680608365019, "grad_norm": 0.3671875, "learning_rate": 9.170529202357705e-05, "loss": 2.2048, "num_input_tokens_seen": 626442240, "step": 4780, "train_runtime": 200924.6962, "train_tokens_per_second": 3117.796 }, { "epoch": 2.9137642585551333, "grad_norm": 0.388671875, "learning_rate": 9.075409210596758e-05, "loss": 2.2148, "num_input_tokens_seen": 627752960, "step": 4790, "train_runtime": 201204.8499, "train_tokens_per_second": 3119.969 }, { "epoch": 2.919847908745247, "grad_norm": 0.37890625, "learning_rate": 8.980675584139152e-05, "loss": 2.1909, "num_input_tokens_seen": 629063680, "step": 4800, "train_runtime": 201490.3533, "train_tokens_per_second": 3122.054 }, { "epoch": 2.9259315589353614, "grad_norm": 0.37890625, "learning_rate": 8.886330621429283e-05, "loss": 2.1946, "num_input_tokens_seen": 630374400, "step": 4810, "train_runtime": 201771.7145, "train_tokens_per_second": 3124.196 }, { "epoch": 2.9320152091254754, "grad_norm": 0.376953125, "learning_rate": 8.792376611481729e-05, "loss": 2.2246, "num_input_tokens_seen": 631685120, "step": 4820, "train_runtime": 202053.0225, "train_tokens_per_second": 3126.333 }, { "epoch": 2.9380988593155895, "grad_norm": 0.388671875, "learning_rate": 8.698815833825654e-05, "loss": 2.2071, "num_input_tokens_seen": 632995840, "step": 4830, "train_runtime": 202335.6315, "train_tokens_per_second": 3128.445 }, { "epoch": 2.9441825095057035, "grad_norm": 0.390625, "learning_rate": 8.605650558449588e-05, "loss": 2.1973, "num_input_tokens_seen": 634306560, "step": 4840, "train_runtime": 202616.803, "train_tokens_per_second": 3130.572 }, { "epoch": 2.9502661596958175, "grad_norm": 0.400390625, "learning_rate": 8.512883045746297e-05, "loss": 2.2143, "num_input_tokens_seen": 635617280, "step": 4850, "train_runtime": 202898.4133, "train_tokens_per_second": 3132.687 }, { "epoch": 2.9563498098859315, "grad_norm": 0.3828125, "learning_rate": 8.420515546457966e-05, "loss": 2.2054, "num_input_tokens_seen": 636928000, "step": 4860, "train_runtime": 203178.4938, "train_tokens_per_second": 3134.82 }, { "epoch": 2.9624334600760456, "grad_norm": 0.373046875, "learning_rate": 8.32855030162157e-05, "loss": 2.2096, "num_input_tokens_seen": 638238720, "step": 4870, "train_runtime": 203457.2711, "train_tokens_per_second": 3136.967 }, { "epoch": 2.9685171102661596, "grad_norm": 0.396484375, "learning_rate": 8.236989542514536e-05, "loss": 2.2334, "num_input_tokens_seen": 639549440, "step": 4880, "train_runtime": 203736.9483, "train_tokens_per_second": 3139.094 }, { "epoch": 2.9746007604562736, "grad_norm": 0.365234375, "learning_rate": 8.145835490600548e-05, "loss": 2.2161, "num_input_tokens_seen": 640860160, "step": 4890, "train_runtime": 204018.1461, "train_tokens_per_second": 3141.192 }, { "epoch": 2.9806844106463877, "grad_norm": 0.4140625, "learning_rate": 8.055090357475711e-05, "loss": 2.2009, "num_input_tokens_seen": 642170880, "step": 4900, "train_runtime": 204300.372, "train_tokens_per_second": 3143.268 }, { "epoch": 2.9867680608365017, "grad_norm": 0.400390625, "learning_rate": 7.964756344814892e-05, "loss": 2.2255, "num_input_tokens_seen": 643481600, "step": 4910, "train_runtime": 204582.9097, "train_tokens_per_second": 3145.334 }, { "epoch": 2.992851711026616, "grad_norm": 0.37109375, "learning_rate": 7.87483564431822e-05, "loss": 2.1963, "num_input_tokens_seen": 644792320, "step": 4920, "train_runtime": 204865.3517, "train_tokens_per_second": 3147.396 }, { "epoch": 2.9989353612167298, "grad_norm": 0.388671875, "learning_rate": 7.78533043765802e-05, "loss": 2.2164, "num_input_tokens_seen": 646103040, "step": 4930, "train_runtime": 205148.8225, "train_tokens_per_second": 3149.436 }, { "epoch": 3.0048669201520912, "grad_norm": 1.1015625, "learning_rate": 7.696242896425806e-05, "loss": 1.8253, "num_input_tokens_seen": 647372800, "step": 4940, "train_runtime": 205422.0038, "train_tokens_per_second": 3151.429 }, { "epoch": 3.0109505703422053, "grad_norm": 0.78125, "learning_rate": 7.607575182079628e-05, "loss": 1.6715, "num_input_tokens_seen": 648683520, "step": 4950, "train_runtime": 205704.4519, "train_tokens_per_second": 3153.473 }, { "epoch": 3.0170342205323193, "grad_norm": 0.5078125, "learning_rate": 7.519329445891612e-05, "loss": 1.657, "num_input_tokens_seen": 649994240, "step": 4960, "train_runtime": 205986.9223, "train_tokens_per_second": 3155.512 }, { "epoch": 3.0231178707224333, "grad_norm": 0.5078125, "learning_rate": 7.431507828895793e-05, "loss": 1.6522, "num_input_tokens_seen": 651304960, "step": 4970, "train_runtime": 206268.5761, "train_tokens_per_second": 3157.558 }, { "epoch": 3.0292015209125474, "grad_norm": 0.45703125, "learning_rate": 7.344112461836108e-05, "loss": 1.6416, "num_input_tokens_seen": 652615680, "step": 4980, "train_runtime": 206547.7316, "train_tokens_per_second": 3159.636 }, { "epoch": 3.0352851711026614, "grad_norm": 0.50390625, "learning_rate": 7.25714546511478e-05, "loss": 1.6381, "num_input_tokens_seen": 653926400, "step": 4990, "train_runtime": 206828.4798, "train_tokens_per_second": 3161.685 }, { "epoch": 3.041368821292776, "grad_norm": 0.52734375, "learning_rate": 7.170608948740825e-05, "loss": 1.6629, "num_input_tokens_seen": 655237120, "step": 5000, "train_runtime": 207110.1126, "train_tokens_per_second": 3163.714 }, { "epoch": 3.04745247148289, "grad_norm": 0.458984375, "learning_rate": 7.084505012278857e-05, "loss": 1.6385, "num_input_tokens_seen": 656547840, "step": 5010, "train_runtime": 207400.0315, "train_tokens_per_second": 3165.611 }, { "epoch": 3.053536121673004, "grad_norm": 0.498046875, "learning_rate": 6.998835744798173e-05, "loss": 1.6357, "num_input_tokens_seen": 657858560, "step": 5020, "train_runtime": 207682.2079, "train_tokens_per_second": 3167.621 }, { "epoch": 3.059619771863118, "grad_norm": 0.52734375, "learning_rate": 6.91360322482202e-05, "loss": 1.6394, "num_input_tokens_seen": 659169280, "step": 5030, "train_runtime": 207963.6755, "train_tokens_per_second": 3169.637 }, { "epoch": 3.065703422053232, "grad_norm": 0.462890625, "learning_rate": 6.82880952027724e-05, "loss": 1.6288, "num_input_tokens_seen": 660480000, "step": 5040, "train_runtime": 208244.7333, "train_tokens_per_second": 3171.653 }, { "epoch": 3.071787072243346, "grad_norm": 0.498046875, "learning_rate": 6.744456688444042e-05, "loss": 1.614, "num_input_tokens_seen": 661790720, "step": 5050, "train_runtime": 208526.526, "train_tokens_per_second": 3173.652 }, { "epoch": 3.07787072243346, "grad_norm": 0.498046875, "learning_rate": 6.660546775906079e-05, "loss": 1.6317, "num_input_tokens_seen": 663101440, "step": 5060, "train_runtime": 208806.0903, "train_tokens_per_second": 3175.681 }, { "epoch": 3.083954372623574, "grad_norm": 0.46484375, "learning_rate": 6.577081818500838e-05, "loss": 1.6632, "num_input_tokens_seen": 664412160, "step": 5070, "train_runtime": 209085.8717, "train_tokens_per_second": 3177.7 }, { "epoch": 3.090038022813688, "grad_norm": 0.4453125, "learning_rate": 6.494063841270223e-05, "loss": 1.6428, "num_input_tokens_seen": 665722880, "step": 5080, "train_runtime": 209364.6273, "train_tokens_per_second": 3179.729 }, { "epoch": 3.096121673003802, "grad_norm": 0.458984375, "learning_rate": 6.411494858411415e-05, "loss": 1.6511, "num_input_tokens_seen": 667033600, "step": 5090, "train_runtime": 209643.9003, "train_tokens_per_second": 3181.746 }, { "epoch": 3.102205323193916, "grad_norm": 0.451171875, "learning_rate": 6.329376873228022e-05, "loss": 1.659, "num_input_tokens_seen": 668344320, "step": 5100, "train_runtime": 209922.9535, "train_tokens_per_second": 3183.76 }, { "epoch": 3.10828897338403, "grad_norm": 0.451171875, "learning_rate": 6.247711878081463e-05, "loss": 1.6451, "num_input_tokens_seen": 669655040, "step": 5110, "train_runtime": 210201.0119, "train_tokens_per_second": 3185.784 }, { "epoch": 3.1143726235741447, "grad_norm": 0.4609375, "learning_rate": 6.16650185434261e-05, "loss": 1.6196, "num_input_tokens_seen": 670965760, "step": 5120, "train_runtime": 210479.9689, "train_tokens_per_second": 3187.789 }, { "epoch": 3.1204562737642587, "grad_norm": 0.453125, "learning_rate": 6.085748772343772e-05, "loss": 1.6467, "num_input_tokens_seen": 672276480, "step": 5130, "train_runtime": 210757.9701, "train_tokens_per_second": 3189.803 }, { "epoch": 3.1265399239543727, "grad_norm": 0.466796875, "learning_rate": 6.005454591330842e-05, "loss": 1.6439, "num_input_tokens_seen": 673587200, "step": 5140, "train_runtime": 211036.4458, "train_tokens_per_second": 3191.805 }, { "epoch": 3.1326235741444868, "grad_norm": 0.5078125, "learning_rate": 5.9256212594157505e-05, "loss": 1.6398, "num_input_tokens_seen": 674897920, "step": 5150, "train_runtime": 211317.0466, "train_tokens_per_second": 3193.769 }, { "epoch": 3.138707224334601, "grad_norm": 0.455078125, "learning_rate": 5.8462507135292524e-05, "loss": 1.6413, "num_input_tokens_seen": 676208640, "step": 5160, "train_runtime": 211596.8045, "train_tokens_per_second": 3195.741 }, { "epoch": 3.144790874524715, "grad_norm": 0.46875, "learning_rate": 5.7673448793738956e-05, "loss": 1.6247, "num_input_tokens_seen": 677519360, "step": 5170, "train_runtime": 211876.5776, "train_tokens_per_second": 3197.708 }, { "epoch": 3.150874524714829, "grad_norm": 0.458984375, "learning_rate": 5.688905671377304e-05, "loss": 1.6543, "num_input_tokens_seen": 678830080, "step": 5180, "train_runtime": 212157.0767, "train_tokens_per_second": 3199.658 }, { "epoch": 3.156958174904943, "grad_norm": 0.474609375, "learning_rate": 5.610934992645736e-05, "loss": 1.6249, "num_input_tokens_seen": 680140800, "step": 5190, "train_runtime": 212435.2324, "train_tokens_per_second": 3201.638 }, { "epoch": 3.163041825095057, "grad_norm": 0.4765625, "learning_rate": 5.533434734917914e-05, "loss": 1.6163, "num_input_tokens_seen": 681451520, "step": 5200, "train_runtime": 212713.6486, "train_tokens_per_second": 3203.61 }, { "epoch": 3.169125475285171, "grad_norm": 0.466796875, "learning_rate": 5.456406778519099e-05, "loss": 1.6824, "num_input_tokens_seen": 682762240, "step": 5210, "train_runtime": 212991.8229, "train_tokens_per_second": 3205.58 }, { "epoch": 3.175209125475285, "grad_norm": 0.486328125, "learning_rate": 5.3798529923155286e-05, "loss": 1.6088, "num_input_tokens_seen": 684072960, "step": 5220, "train_runtime": 213269.9689, "train_tokens_per_second": 3207.545 }, { "epoch": 3.1812927756653995, "grad_norm": 0.447265625, "learning_rate": 5.303775233669003e-05, "loss": 1.6733, "num_input_tokens_seen": 685383680, "step": 5230, "train_runtime": 213547.7284, "train_tokens_per_second": 3209.511 }, { "epoch": 3.1873764258555135, "grad_norm": 0.47265625, "learning_rate": 5.22817534839187e-05, "loss": 1.648, "num_input_tokens_seen": 686694400, "step": 5240, "train_runtime": 213827.0794, "train_tokens_per_second": 3211.447 }, { "epoch": 3.1934600760456275, "grad_norm": 0.46875, "learning_rate": 5.153055170702231e-05, "loss": 1.6205, "num_input_tokens_seen": 688005120, "step": 5250, "train_runtime": 214106.7604, "train_tokens_per_second": 3213.374 }, { "epoch": 3.1995437262357416, "grad_norm": 0.453125, "learning_rate": 5.078416523179427e-05, "loss": 1.6327, "num_input_tokens_seen": 689315840, "step": 5260, "train_runtime": 214384.661, "train_tokens_per_second": 3215.323 }, { "epoch": 3.2056273764258556, "grad_norm": 0.455078125, "learning_rate": 5.00426121671983e-05, "loss": 1.6355, "num_input_tokens_seen": 690626560, "step": 5270, "train_runtime": 214663.7036, "train_tokens_per_second": 3217.249 }, { "epoch": 3.2117110266159696, "grad_norm": 0.48046875, "learning_rate": 4.930591050492913e-05, "loss": 1.6226, "num_input_tokens_seen": 691937280, "step": 5280, "train_runtime": 214941.4149, "train_tokens_per_second": 3219.19 }, { "epoch": 3.2177946768060837, "grad_norm": 0.453125, "learning_rate": 4.857407811897566e-05, "loss": 1.6398, "num_input_tokens_seen": 693248000, "step": 5290, "train_runtime": 215219.1478, "train_tokens_per_second": 3221.126 }, { "epoch": 3.2238783269961977, "grad_norm": 0.453125, "learning_rate": 4.7847132765187676e-05, "loss": 1.6255, "num_input_tokens_seen": 694558720, "step": 5300, "train_runtime": 215497.1854, "train_tokens_per_second": 3223.052 }, { "epoch": 3.2299619771863117, "grad_norm": 0.4453125, "learning_rate": 4.712509208084506e-05, "loss": 1.6393, "num_input_tokens_seen": 695869440, "step": 5310, "train_runtime": 215774.8903, "train_tokens_per_second": 3224.979 }, { "epoch": 3.2360456273764258, "grad_norm": 0.462890625, "learning_rate": 4.640797358422938e-05, "loss": 1.6536, "num_input_tokens_seen": 697180160, "step": 5320, "train_runtime": 216052.2699, "train_tokens_per_second": 3226.905 }, { "epoch": 3.24212927756654, "grad_norm": 0.45703125, "learning_rate": 4.569579467419943e-05, "loss": 1.6502, "num_input_tokens_seen": 698490880, "step": 5330, "train_runtime": 216334.0737, "train_tokens_per_second": 3228.76 }, { "epoch": 3.248212927756654, "grad_norm": 0.455078125, "learning_rate": 4.49885726297688e-05, "loss": 1.6125, "num_input_tokens_seen": 699801600, "step": 5340, "train_runtime": 216611.371, "train_tokens_per_second": 3230.678 }, { "epoch": 3.254296577946768, "grad_norm": 0.466796875, "learning_rate": 4.428632460968671e-05, "loss": 1.6049, "num_input_tokens_seen": 701112320, "step": 5350, "train_runtime": 216889.4967, "train_tokens_per_second": 3232.578 }, { "epoch": 3.2603802281368823, "grad_norm": 0.466796875, "learning_rate": 4.358906765202167e-05, "loss": 1.6302, "num_input_tokens_seen": 702423040, "step": 5360, "train_runtime": 217167.1521, "train_tokens_per_second": 3234.481 }, { "epoch": 3.2664638783269964, "grad_norm": 0.453125, "learning_rate": 4.2896818673748256e-05, "loss": 1.6092, "num_input_tokens_seen": 703733760, "step": 5370, "train_runtime": 217444.6002, "train_tokens_per_second": 3236.382 }, { "epoch": 3.2725475285171104, "grad_norm": 0.48046875, "learning_rate": 4.220959447033629e-05, "loss": 1.668, "num_input_tokens_seen": 705044480, "step": 5380, "train_runtime": 217722.1571, "train_tokens_per_second": 3238.276 }, { "epoch": 3.2786311787072244, "grad_norm": 0.486328125, "learning_rate": 4.1527411715343774e-05, "loss": 1.6002, "num_input_tokens_seen": 706355200, "step": 5390, "train_runtime": 218000.8989, "train_tokens_per_second": 3240.148 }, { "epoch": 3.2847148288973385, "grad_norm": 0.447265625, "learning_rate": 4.0850286960012335e-05, "loss": 1.6425, "num_input_tokens_seen": 707665920, "step": 5400, "train_runtime": 218279.0655, "train_tokens_per_second": 3242.024 }, { "epoch": 3.2907984790874525, "grad_norm": 0.453125, "learning_rate": 4.01782366328651e-05, "loss": 1.6527, "num_input_tokens_seen": 708976640, "step": 5410, "train_runtime": 218558.3443, "train_tokens_per_second": 3243.878 }, { "epoch": 3.2968821292775665, "grad_norm": 0.45703125, "learning_rate": 3.951127703930882e-05, "loss": 1.6389, "num_input_tokens_seen": 710287360, "step": 5420, "train_runtime": 218838.5125, "train_tokens_per_second": 3245.715 }, { "epoch": 3.3029657794676806, "grad_norm": 0.455078125, "learning_rate": 3.884942436123784e-05, "loss": 1.625, "num_input_tokens_seen": 711598080, "step": 5430, "train_runtime": 219119.7533, "train_tokens_per_second": 3247.53 }, { "epoch": 3.3090494296577946, "grad_norm": 0.4453125, "learning_rate": 3.819269465664157e-05, "loss": 1.6137, "num_input_tokens_seen": 712908800, "step": 5440, "train_runtime": 219397.4568, "train_tokens_per_second": 3249.394 }, { "epoch": 3.3151330798479086, "grad_norm": 0.439453125, "learning_rate": 3.754110385921502e-05, "loss": 1.6283, "num_input_tokens_seen": 714219520, "step": 5450, "train_runtime": 219677.914, "train_tokens_per_second": 3251.212 }, { "epoch": 3.3212167300380226, "grad_norm": 0.466796875, "learning_rate": 3.68946677779719e-05, "loss": 1.6457, "num_input_tokens_seen": 715530240, "step": 5460, "train_runtime": 219958.1204, "train_tokens_per_second": 3253.029 }, { "epoch": 3.327300380228137, "grad_norm": 0.447265625, "learning_rate": 3.6253402096861456e-05, "loss": 1.6237, "num_input_tokens_seen": 716840960, "step": 5470, "train_runtime": 220239.0455, "train_tokens_per_second": 3254.831 }, { "epoch": 3.3333840304182507, "grad_norm": 0.458984375, "learning_rate": 3.56173223743877e-05, "loss": 1.6404, "num_input_tokens_seen": 718151680, "step": 5480, "train_runtime": 220519.7735, "train_tokens_per_second": 3256.632 }, { "epoch": 3.339467680608365, "grad_norm": 0.46875, "learning_rate": 3.4986444043231926e-05, "loss": 1.6324, "num_input_tokens_seen": 719462400, "step": 5490, "train_runtime": 220800.5642, "train_tokens_per_second": 3258.426 }, { "epoch": 3.345551330798479, "grad_norm": 0.474609375, "learning_rate": 3.436078240987836e-05, "loss": 1.6478, "num_input_tokens_seen": 720773120, "step": 5500, "train_runtime": 221081.8118, "train_tokens_per_second": 3260.21 }, { "epoch": 3.3516349809885932, "grad_norm": 0.453125, "learning_rate": 3.3740352654242856e-05, "loss": 1.6567, "num_input_tokens_seen": 722083840, "step": 5510, "train_runtime": 221367.5106, "train_tokens_per_second": 3261.923 }, { "epoch": 3.3577186311787073, "grad_norm": 0.458984375, "learning_rate": 3.312516982930422e-05, "loss": 1.6376, "num_input_tokens_seen": 723394560, "step": 5520, "train_runtime": 221645.3627, "train_tokens_per_second": 3263.748 }, { "epoch": 3.3638022813688213, "grad_norm": 0.44140625, "learning_rate": 3.251524886073964e-05, "loss": 1.6641, "num_input_tokens_seen": 724705280, "step": 5530, "train_runtime": 221924.8737, "train_tokens_per_second": 3265.543 }, { "epoch": 3.3698859315589353, "grad_norm": 0.46484375, "learning_rate": 3.191060454656205e-05, "loss": 1.6531, "num_input_tokens_seen": 726016000, "step": 5540, "train_runtime": 222203.4442, "train_tokens_per_second": 3267.348 }, { "epoch": 3.3759695817490494, "grad_norm": 0.455078125, "learning_rate": 3.1311251556761064e-05, "loss": 1.6111, "num_input_tokens_seen": 727326720, "step": 5550, "train_runtime": 222483.2619, "train_tokens_per_second": 3269.13 }, { "epoch": 3.3820532319391634, "grad_norm": 0.466796875, "learning_rate": 3.071720443294748e-05, "loss": 1.6169, "num_input_tokens_seen": 728637440, "step": 5560, "train_runtime": 222764.1882, "train_tokens_per_second": 3270.891 }, { "epoch": 3.3881368821292774, "grad_norm": 0.46875, "learning_rate": 3.0128477588000052e-05, "loss": 1.6792, "num_input_tokens_seen": 729948160, "step": 5570, "train_runtime": 223046.4055, "train_tokens_per_second": 3272.629 }, { "epoch": 3.3942205323193915, "grad_norm": 0.447265625, "learning_rate": 2.954508530571609e-05, "loss": 1.6474, "num_input_tokens_seen": 731258880, "step": 5580, "train_runtime": 223331.3629, "train_tokens_per_second": 3274.322 }, { "epoch": 3.4003041825095055, "grad_norm": 0.4609375, "learning_rate": 2.8967041740464633e-05, "loss": 1.6186, "num_input_tokens_seen": 732569600, "step": 5590, "train_runtime": 223620.5675, "train_tokens_per_second": 3275.949 }, { "epoch": 3.40638783269962, "grad_norm": 0.44921875, "learning_rate": 2.8394360916843388e-05, "loss": 1.6077, "num_input_tokens_seen": 733880320, "step": 5600, "train_runtime": 224039.407, "train_tokens_per_second": 3275.675 }, { "epoch": 3.412471482889734, "grad_norm": 0.451171875, "learning_rate": 2.782705672933794e-05, "loss": 1.6165, "num_input_tokens_seen": 735191040, "step": 5610, "train_runtime": 224502.8879, "train_tokens_per_second": 3274.751 }, { "epoch": 3.418555133079848, "grad_norm": 0.4609375, "learning_rate": 2.7265142941985316e-05, "loss": 1.6244, "num_input_tokens_seen": 736501760, "step": 5620, "train_runtime": 224972.3357, "train_tokens_per_second": 3273.744 }, { "epoch": 3.424638783269962, "grad_norm": 0.482421875, "learning_rate": 2.6708633188039455e-05, "loss": 1.6326, "num_input_tokens_seen": 737812480, "step": 5630, "train_runtime": 225440.8866, "train_tokens_per_second": 3272.754 }, { "epoch": 3.430722433460076, "grad_norm": 0.453125, "learning_rate": 2.6157540969640652e-05, "loss": 1.6292, "num_input_tokens_seen": 739123200, "step": 5640, "train_runtime": 225911.9651, "train_tokens_per_second": 3271.731 }, { "epoch": 3.43680608365019, "grad_norm": 0.455078125, "learning_rate": 2.5611879657488096e-05, "loss": 1.6367, "num_input_tokens_seen": 740433920, "step": 5650, "train_runtime": 226382.517, "train_tokens_per_second": 3270.72 }, { "epoch": 3.442889733840304, "grad_norm": 0.494140625, "learning_rate": 2.5071662490515233e-05, "loss": 1.6283, "num_input_tokens_seen": 741744640, "step": 5660, "train_runtime": 226853.4792, "train_tokens_per_second": 3269.708 }, { "epoch": 3.448973384030418, "grad_norm": 0.451171875, "learning_rate": 2.4536902575568758e-05, "loss": 1.642, "num_input_tokens_seen": 743055360, "step": 5670, "train_runtime": 227323.0676, "train_tokens_per_second": 3268.72 }, { "epoch": 3.4550570342205322, "grad_norm": 0.4765625, "learning_rate": 2.4007612887090524e-05, "loss": 1.6326, "num_input_tokens_seen": 744366080, "step": 5680, "train_runtime": 227792.7324, "train_tokens_per_second": 3267.734 }, { "epoch": 3.4611406844106463, "grad_norm": 0.443359375, "learning_rate": 2.3483806266802632e-05, "loss": 1.6121, "num_input_tokens_seen": 745676800, "step": 5690, "train_runtime": 228265.1981, "train_tokens_per_second": 3266.713 }, { "epoch": 3.4672243346007603, "grad_norm": 0.46875, "learning_rate": 2.2965495423396126e-05, "loss": 1.6207, "num_input_tokens_seen": 746987520, "step": 5700, "train_runtime": 228741.9712, "train_tokens_per_second": 3265.634 }, { "epoch": 3.4733079847908748, "grad_norm": 0.462890625, "learning_rate": 2.2452692932222575e-05, "loss": 1.6266, "num_input_tokens_seen": 748298240, "step": 5710, "train_runtime": 229218.7728, "train_tokens_per_second": 3264.559 }, { "epoch": 3.4793916349809884, "grad_norm": 0.453125, "learning_rate": 2.194541123498872e-05, "loss": 1.6278, "num_input_tokens_seen": 749608960, "step": 5720, "train_runtime": 229696.1749, "train_tokens_per_second": 3263.48 }, { "epoch": 3.485475285171103, "grad_norm": 0.439453125, "learning_rate": 2.144366263945488e-05, "loss": 1.6114, "num_input_tokens_seen": 750919680, "step": 5730, "train_runtime": 230173.4943, "train_tokens_per_second": 3262.407 }, { "epoch": 3.491558935361217, "grad_norm": 0.44140625, "learning_rate": 2.094745931913633e-05, "loss": 1.6453, "num_input_tokens_seen": 752230400, "step": 5740, "train_runtime": 230651.4755, "train_tokens_per_second": 3261.329 }, { "epoch": 3.497642585551331, "grad_norm": 0.451171875, "learning_rate": 2.0456813313007777e-05, "loss": 1.6273, "num_input_tokens_seen": 753541120, "step": 5750, "train_runtime": 231130.0952, "train_tokens_per_second": 3260.247 }, { "epoch": 3.503726235741445, "grad_norm": 0.453125, "learning_rate": 1.997173652521142e-05, "loss": 1.6474, "num_input_tokens_seen": 754851840, "step": 5760, "train_runtime": 231605.5777, "train_tokens_per_second": 3259.213 }, { "epoch": 3.509809885931559, "grad_norm": 0.453125, "learning_rate": 1.9492240724768128e-05, "loss": 1.6206, "num_input_tokens_seen": 756162560, "step": 5770, "train_runtime": 232080.7709, "train_tokens_per_second": 3258.187 }, { "epoch": 3.515893536121673, "grad_norm": 0.4453125, "learning_rate": 1.9018337545291668e-05, "loss": 1.6258, "num_input_tokens_seen": 757473280, "step": 5780, "train_runtime": 232556.6923, "train_tokens_per_second": 3257.155 }, { "epoch": 3.521977186311787, "grad_norm": 0.451171875, "learning_rate": 1.8550038484706738e-05, "loss": 1.6021, "num_input_tokens_seen": 758784000, "step": 5790, "train_runtime": 233033.1187, "train_tokens_per_second": 3256.121 }, { "epoch": 3.528060836501901, "grad_norm": 0.462890625, "learning_rate": 1.8087354904969987e-05, "loss": 1.6396, "num_input_tokens_seen": 760094720, "step": 5800, "train_runtime": 233507.6875, "train_tokens_per_second": 3255.116 }, { "epoch": 3.534144486692015, "grad_norm": 0.4375, "learning_rate": 1.7630298031794084e-05, "loss": 1.6261, "num_input_tokens_seen": 761405440, "step": 5810, "train_runtime": 233983.9121, "train_tokens_per_second": 3254.093 }, { "epoch": 3.540228136882129, "grad_norm": 0.447265625, "learning_rate": 1.717887895437559e-05, "loss": 1.6206, "num_input_tokens_seen": 762716160, "step": 5820, "train_runtime": 234460.178, "train_tokens_per_second": 3253.073 }, { "epoch": 3.546311787072243, "grad_norm": 0.435546875, "learning_rate": 1.6733108625125866e-05, "loss": 1.6487, "num_input_tokens_seen": 764026880, "step": 5830, "train_runtime": 234938.1446, "train_tokens_per_second": 3252.034 }, { "epoch": 3.5523954372623576, "grad_norm": 0.45703125, "learning_rate": 1.62929978594053e-05, "loss": 1.6688, "num_input_tokens_seen": 765337600, "step": 5840, "train_runtime": 235413.0674, "train_tokens_per_second": 3251.041 }, { "epoch": 3.558479087452471, "grad_norm": 0.431640625, "learning_rate": 1.5858557335260926e-05, "loss": 1.6187, "num_input_tokens_seen": 766648320, "step": 5850, "train_runtime": 235888.1153, "train_tokens_per_second": 3250.051 }, { "epoch": 3.5645627376425857, "grad_norm": 0.4296875, "learning_rate": 1.5429797593167372e-05, "loss": 1.5958, "num_input_tokens_seen": 767959040, "step": 5860, "train_runtime": 236366.2531, "train_tokens_per_second": 3249.022 }, { "epoch": 3.5706463878326997, "grad_norm": 0.455078125, "learning_rate": 1.5006729035771078e-05, "loss": 1.6523, "num_input_tokens_seen": 769269760, "step": 5870, "train_runtime": 236845.1897, "train_tokens_per_second": 3247.986 }, { "epoch": 3.5767300380228138, "grad_norm": 0.474609375, "learning_rate": 1.4589361927637907e-05, "loss": 1.6243, "num_input_tokens_seen": 770580480, "step": 5880, "train_runtime": 237322.8914, "train_tokens_per_second": 3246.971 }, { "epoch": 3.582813688212928, "grad_norm": 0.43359375, "learning_rate": 1.417770639500418e-05, "loss": 1.6093, "num_input_tokens_seen": 771891200, "step": 5890, "train_runtime": 237801.8938, "train_tokens_per_second": 3245.942 }, { "epoch": 3.588897338403042, "grad_norm": 0.44140625, "learning_rate": 1.3771772425530932e-05, "loss": 1.6162, "num_input_tokens_seen": 773201920, "step": 5900, "train_runtime": 238279.7947, "train_tokens_per_second": 3244.933 }, { "epoch": 3.594980988593156, "grad_norm": 0.4453125, "learning_rate": 1.3371569868061622e-05, "loss": 1.6657, "num_input_tokens_seen": 774512640, "step": 5910, "train_runtime": 238759.5079, "train_tokens_per_second": 3243.903 }, { "epoch": 3.60106463878327, "grad_norm": 0.43359375, "learning_rate": 1.2977108432383006e-05, "loss": 1.5992, "num_input_tokens_seen": 775823360, "step": 5920, "train_runtime": 239239.9149, "train_tokens_per_second": 3242.868 }, { "epoch": 3.607148288973384, "grad_norm": 0.4375, "learning_rate": 1.2588397688989923e-05, "loss": 1.6096, "num_input_tokens_seen": 777134080, "step": 5930, "train_runtime": 239721.0782, "train_tokens_per_second": 3241.826 }, { "epoch": 3.613231939163498, "grad_norm": 0.439453125, "learning_rate": 1.2205447068852766e-05, "loss": 1.6149, "num_input_tokens_seen": 778444800, "step": 5940, "train_runtime": 240200.8963, "train_tokens_per_second": 3240.807 }, { "epoch": 3.6193155893536124, "grad_norm": 0.427734375, "learning_rate": 1.1828265863188692e-05, "loss": 1.6129, "num_input_tokens_seen": 779755520, "step": 5950, "train_runtime": 240678.7967, "train_tokens_per_second": 3239.818 }, { "epoch": 3.625399239543726, "grad_norm": 0.43359375, "learning_rate": 1.145686322323644e-05, "loss": 1.6548, "num_input_tokens_seen": 781066240, "step": 5960, "train_runtime": 241156.9834, "train_tokens_per_second": 3238.829 }, { "epoch": 3.6314828897338405, "grad_norm": 0.439453125, "learning_rate": 1.109124816003404e-05, "loss": 1.6589, "num_input_tokens_seen": 782376960, "step": 5970, "train_runtime": 241634.9202, "train_tokens_per_second": 3237.847 }, { "epoch": 3.6375665399239545, "grad_norm": 0.439453125, "learning_rate": 1.0731429544200311e-05, "loss": 1.667, "num_input_tokens_seen": 783687680, "step": 5980, "train_runtime": 242110.8067, "train_tokens_per_second": 3236.897 }, { "epoch": 3.6436501901140685, "grad_norm": 0.4375, "learning_rate": 1.0377416105719628e-05, "loss": 1.6387, "num_input_tokens_seen": 784998400, "step": 5990, "train_runtime": 242587.3722, "train_tokens_per_second": 3235.941 }, { "epoch": 3.6497338403041826, "grad_norm": 0.4453125, "learning_rate": 1.0029216433730115e-05, "loss": 1.6584, "num_input_tokens_seen": 786309120, "step": 6000, "train_runtime": 243064.283, "train_tokens_per_second": 3234.984 }, { "epoch": 3.6558174904942966, "grad_norm": 0.4453125, "learning_rate": 9.686838976315172e-06, "loss": 1.6483, "num_input_tokens_seen": 787619840, "step": 6010, "train_runtime": 243540.2153, "train_tokens_per_second": 3234.044 }, { "epoch": 3.6619011406844106, "grad_norm": 0.46484375, "learning_rate": 9.350292040298664e-06, "loss": 1.6513, "num_input_tokens_seen": 788930560, "step": 6020, "train_runtime": 244016.4696, "train_tokens_per_second": 3233.104 }, { "epoch": 3.6679847908745247, "grad_norm": 0.4453125, "learning_rate": 9.019583791043284e-06, "loss": 1.6027, "num_input_tokens_seen": 790241280, "step": 6030, "train_runtime": 244495.7045, "train_tokens_per_second": 3232.127 }, { "epoch": 3.6740684410646387, "grad_norm": 0.453125, "learning_rate": 8.694722252252346e-06, "loss": 1.618, "num_input_tokens_seen": 791552000, "step": 6040, "train_runtime": 244974.1179, "train_tokens_per_second": 3231.166 }, { "epoch": 3.6801520912547527, "grad_norm": 0.45703125, "learning_rate": 8.375715305775327e-06, "loss": 1.6454, "num_input_tokens_seen": 792862720, "step": 6050, "train_runtime": 245450.5375, "train_tokens_per_second": 3230.234 }, { "epoch": 3.6862357414448668, "grad_norm": 0.44140625, "learning_rate": 8.06257069141647e-06, "loss": 1.5994, "num_input_tokens_seen": 794173440, "step": 6060, "train_runtime": 245927.9251, "train_tokens_per_second": 3229.293 }, { "epoch": 3.692319391634981, "grad_norm": 0.431640625, "learning_rate": 7.755296006747098e-06, "loss": 1.6254, "num_input_tokens_seen": 795484160, "step": 6070, "train_runtime": 246405.2565, "train_tokens_per_second": 3228.357 }, { "epoch": 3.6984030418250953, "grad_norm": 0.443359375, "learning_rate": 7.453898706921203e-06, "loss": 1.6246, "num_input_tokens_seen": 796794880, "step": 6080, "train_runtime": 246883.6093, "train_tokens_per_second": 3227.411 }, { "epoch": 3.704486692015209, "grad_norm": 0.44140625, "learning_rate": 7.1583861044946515e-06, "loss": 1.6219, "num_input_tokens_seen": 798105600, "step": 6090, "train_runtime": 247363.466, "train_tokens_per_second": 3226.449 }, { "epoch": 3.7105703422053233, "grad_norm": 0.44140625, "learning_rate": 6.868765369247604e-06, "loss": 1.6234, "num_input_tokens_seen": 799416320, "step": 6100, "train_runtime": 247842.3464, "train_tokens_per_second": 3225.503 }, { "epoch": 3.7166539923954374, "grad_norm": 0.470703125, "learning_rate": 6.585043528010759e-06, "loss": 1.6664, "num_input_tokens_seen": 800727040, "step": 6110, "train_runtime": 248320.6683, "train_tokens_per_second": 3224.569 }, { "epoch": 3.7227376425855514, "grad_norm": 0.43359375, "learning_rate": 6.30722746449483e-06, "loss": 1.6222, "num_input_tokens_seen": 802037760, "step": 6120, "train_runtime": 248798.6552, "train_tokens_per_second": 3223.642 }, { "epoch": 3.7288212927756654, "grad_norm": 0.435546875, "learning_rate": 6.035323919123342e-06, "loss": 1.605, "num_input_tokens_seen": 803348480, "step": 6130, "train_runtime": 249277.2337, "train_tokens_per_second": 3222.711 }, { "epoch": 3.7349049429657795, "grad_norm": 0.439453125, "learning_rate": 5.769339488869374e-06, "loss": 1.6482, "num_input_tokens_seen": 804659200, "step": 6140, "train_runtime": 249757.3827, "train_tokens_per_second": 3221.763 }, { "epoch": 3.7409885931558935, "grad_norm": 0.443359375, "learning_rate": 5.509280627095298e-06, "loss": 1.6258, "num_input_tokens_seen": 805969920, "step": 6150, "train_runtime": 250236.5617, "train_tokens_per_second": 3220.832 }, { "epoch": 3.7470722433460075, "grad_norm": 0.427734375, "learning_rate": 5.2551536433962655e-06, "loss": 1.6281, "num_input_tokens_seen": 807280640, "step": 6160, "train_runtime": 250715.0321, "train_tokens_per_second": 3219.913 }, { "epoch": 3.7531558935361216, "grad_norm": 0.4609375, "learning_rate": 5.006964703447137e-06, "loss": 1.6278, "num_input_tokens_seen": 808591360, "step": 6170, "train_runtime": 251193.7755, "train_tokens_per_second": 3218.994 }, { "epoch": 3.7592395437262356, "grad_norm": 0.453125, "learning_rate": 4.764719828852876e-06, "loss": 1.6374, "num_input_tokens_seen": 809902080, "step": 6180, "train_runtime": 251672.2591, "train_tokens_per_second": 3218.082 }, { "epoch": 3.76532319391635, "grad_norm": 0.43359375, "learning_rate": 4.5284248970025035e-06, "loss": 1.627, "num_input_tokens_seen": 811212800, "step": 6190, "train_runtime": 252151.617, "train_tokens_per_second": 3217.163 }, { "epoch": 3.7714068441064637, "grad_norm": 0.44921875, "learning_rate": 4.2980856409263466e-06, "loss": 1.6303, "num_input_tokens_seen": 812523520, "step": 6200, "train_runtime": 252628.2077, "train_tokens_per_second": 3216.282 }, { "epoch": 3.777490494296578, "grad_norm": 0.435546875, "learning_rate": 4.073707649157182e-06, "loss": 1.6597, "num_input_tokens_seen": 813834240, "step": 6210, "train_runtime": 253105.2476, "train_tokens_per_second": 3215.399 }, { "epoch": 3.783574144486692, "grad_norm": 0.43359375, "learning_rate": 3.855296365594424e-06, "loss": 1.5916, "num_input_tokens_seen": 815144960, "step": 6220, "train_runtime": 253583.0587, "train_tokens_per_second": 3214.509 }, { "epoch": 3.789657794676806, "grad_norm": 0.443359375, "learning_rate": 3.642857089372148e-06, "loss": 1.6516, "num_input_tokens_seen": 816455680, "step": 6230, "train_runtime": 254061.1301, "train_tokens_per_second": 3213.619 }, { "epoch": 3.7957414448669202, "grad_norm": 0.4453125, "learning_rate": 3.436394974730528e-06, "loss": 1.6169, "num_input_tokens_seen": 817766400, "step": 6240, "train_runtime": 254539.0149, "train_tokens_per_second": 3212.735 }, { "epoch": 3.8018250950570343, "grad_norm": 0.4296875, "learning_rate": 3.23591503089074e-06, "loss": 1.6236, "num_input_tokens_seen": 819077120, "step": 6250, "train_runtime": 255018.4108, "train_tokens_per_second": 3211.835 }, { "epoch": 3.8079087452471483, "grad_norm": 0.427734375, "learning_rate": 3.041422121933507e-06, "loss": 1.6347, "num_input_tokens_seen": 820387840, "step": 6260, "train_runtime": 255495.0232, "train_tokens_per_second": 3210.974 }, { "epoch": 3.8139923954372623, "grad_norm": 0.447265625, "learning_rate": 2.852920966680883e-06, "loss": 1.6444, "num_input_tokens_seen": 821698560, "step": 6270, "train_runtime": 255977.0025, "train_tokens_per_second": 3210.048 }, { "epoch": 3.8200760456273763, "grad_norm": 0.4453125, "learning_rate": 2.670416138582099e-06, "loss": 1.657, "num_input_tokens_seen": 823009280, "step": 6280, "train_runtime": 256458.9092, "train_tokens_per_second": 3209.127 }, { "epoch": 3.8261596958174904, "grad_norm": 0.4296875, "learning_rate": 2.4939120656022607e-06, "loss": 1.6014, "num_input_tokens_seen": 824320000, "step": 6290, "train_runtime": 256940.8942, "train_tokens_per_second": 3208.209 }, { "epoch": 3.8322433460076044, "grad_norm": 0.453125, "learning_rate": 2.3234130301151568e-06, "loss": 1.6304, "num_input_tokens_seen": 825630720, "step": 6300, "train_runtime": 257422.8169, "train_tokens_per_second": 3207.294 }, { "epoch": 3.8383269961977184, "grad_norm": 0.4296875, "learning_rate": 2.158923168799232e-06, "loss": 1.6337, "num_input_tokens_seen": 826941440, "step": 6310, "train_runtime": 257901.9133, "train_tokens_per_second": 3206.418 }, { "epoch": 3.844410646387833, "grad_norm": 0.435546875, "learning_rate": 2.0004464725372763e-06, "loss": 1.6302, "num_input_tokens_seen": 828252160, "step": 6320, "train_runtime": 258382.2286, "train_tokens_per_second": 3205.531 }, { "epoch": 3.8504942965779465, "grad_norm": 0.439453125, "learning_rate": 1.8479867863195333e-06, "loss": 1.6333, "num_input_tokens_seen": 829562880, "step": 6330, "train_runtime": 258861.7851, "train_tokens_per_second": 3204.656 }, { "epoch": 3.856577946768061, "grad_norm": 0.46484375, "learning_rate": 1.701547809150522e-06, "loss": 1.6353, "num_input_tokens_seen": 830873600, "step": 6340, "train_runtime": 259339.8156, "train_tokens_per_second": 3203.803 }, { "epoch": 3.862661596958175, "grad_norm": 0.4375, "learning_rate": 1.5611330939592494e-06, "loss": 1.6396, "num_input_tokens_seen": 832184320, "step": 6350, "train_runtime": 259817.9997, "train_tokens_per_second": 3202.951 }, { "epoch": 3.868745247148289, "grad_norm": 0.453125, "learning_rate": 1.4267460475128614e-06, "loss": 1.6672, "num_input_tokens_seen": 833495040, "step": 6360, "train_runtime": 260295.7867, "train_tokens_per_second": 3202.107 }, { "epoch": 3.874828897338403, "grad_norm": 0.4375, "learning_rate": 1.298389930334265e-06, "loss": 1.6369, "num_input_tokens_seen": 834805760, "step": 6370, "train_runtime": 260772.7931, "train_tokens_per_second": 3201.276 }, { "epoch": 3.880912547528517, "grad_norm": 0.43359375, "learning_rate": 1.1760678566227479e-06, "loss": 1.61, "num_input_tokens_seen": 836116480, "step": 6380, "train_runtime": 261249.3034, "train_tokens_per_second": 3200.454 }, { "epoch": 3.886996197718631, "grad_norm": 0.431640625, "learning_rate": 1.0597827941786486e-06, "loss": 1.6572, "num_input_tokens_seen": 837427200, "step": 6390, "train_runtime": 261725.5029, "train_tokens_per_second": 3199.639 }, { "epoch": 3.893079847908745, "grad_norm": 0.435546875, "learning_rate": 9.495375643311655e-07, "loss": 1.6225, "num_input_tokens_seen": 838737920, "step": 6400, "train_runtime": 262202.5532, "train_tokens_per_second": 3198.817 }, { "epoch": 3.899163498098859, "grad_norm": 0.427734375, "learning_rate": 8.453348418700213e-07, "loss": 1.6146, "num_input_tokens_seen": 840048640, "step": 6410, "train_runtime": 262679.9257, "train_tokens_per_second": 3197.993 }, { "epoch": 3.9052471482889732, "grad_norm": 0.447265625, "learning_rate": 7.471771549805162e-07, "loss": 1.6213, "num_input_tokens_seen": 841359360, "step": 6420, "train_runtime": 263158.6243, "train_tokens_per_second": 3197.157 }, { "epoch": 3.9113307984790877, "grad_norm": 0.447265625, "learning_rate": 6.55066885182215e-07, "loss": 1.6522, "num_input_tokens_seen": 842670080, "step": 6430, "train_runtime": 263636.9111, "train_tokens_per_second": 3196.328 }, { "epoch": 3.9174144486692013, "grad_norm": 0.4375, "learning_rate": 5.690062672711605e-07, "loss": 1.6368, "num_input_tokens_seen": 843980800, "step": 6440, "train_runtime": 264114.9361, "train_tokens_per_second": 3195.506 }, { "epoch": 3.9234980988593158, "grad_norm": 0.44140625, "learning_rate": 4.889973892656107e-07, "loss": 1.6258, "num_input_tokens_seen": 845291520, "step": 6450, "train_runtime": 264593.569, "train_tokens_per_second": 3194.679 }, { "epoch": 3.92958174904943, "grad_norm": 0.439453125, "learning_rate": 4.1504219235546923e-07, "loss": 1.6317, "num_input_tokens_seen": 846602240, "step": 6460, "train_runtime": 265072.6691, "train_tokens_per_second": 3193.85 }, { "epoch": 3.935665399239544, "grad_norm": 0.41796875, "learning_rate": 3.471424708551274e-07, "loss": 1.6442, "num_input_tokens_seen": 847912960, "step": 6470, "train_runtime": 265551.9975, "train_tokens_per_second": 3193.02 }, { "epoch": 3.941749049429658, "grad_norm": 0.44140625, "learning_rate": 2.8529987215986096e-07, "loss": 1.6448, "num_input_tokens_seen": 849223680, "step": 6480, "train_runtime": 266031.0263, "train_tokens_per_second": 3192.198 }, { "epoch": 3.947832699619772, "grad_norm": 0.44140625, "learning_rate": 2.295158967060007e-07, "loss": 1.6409, "num_input_tokens_seen": 850534400, "step": 6490, "train_runtime": 266511.6606, "train_tokens_per_second": 3191.359 }, { "epoch": 3.953916349809886, "grad_norm": 0.431640625, "learning_rate": 1.7979189793448925e-07, "loss": 1.6319, "num_input_tokens_seen": 851845120, "step": 6500, "train_runtime": 266991.1154, "train_tokens_per_second": 3190.537 }, { "epoch": 3.96, "grad_norm": 0.478515625, "learning_rate": 1.3612908225796306e-07, "loss": 1.6569, "num_input_tokens_seen": 853155840, "step": 6510, "train_runtime": 267470.5957, "train_tokens_per_second": 3189.718 }, { "epoch": 3.966083650190114, "grad_norm": 0.427734375, "learning_rate": 9.852850903155353e-08, "loss": 1.6181, "num_input_tokens_seen": 854466560, "step": 6520, "train_runtime": 267949.2792, "train_tokens_per_second": 3188.912 }, { "epoch": 3.972167300380228, "grad_norm": 0.439453125, "learning_rate": 6.699109052715758e-08, "loss": 1.6236, "num_input_tokens_seen": 855777280, "step": 6530, "train_runtime": 268429.2552, "train_tokens_per_second": 3188.092 }, { "epoch": 3.978250950570342, "grad_norm": 0.435546875, "learning_rate": 4.151759191137194e-08, "loss": 1.6233, "num_input_tokens_seen": 857088000, "step": 6540, "train_runtime": 268908.9629, "train_tokens_per_second": 3187.279 }, { "epoch": 3.984334600760456, "grad_norm": 0.439453125, "learning_rate": 2.210863122678597e-08, "loss": 1.6551, "num_input_tokens_seen": 858398720, "step": 6550, "train_runtime": 269388.3681, "train_tokens_per_second": 3186.473 }, { "epoch": 3.9904182509505706, "grad_norm": 0.431640625, "learning_rate": 8.764679377132324e-09, "loss": 1.6475, "num_input_tokens_seen": 859709440, "step": 6560, "train_runtime": 269866.3772, "train_tokens_per_second": 3185.686 }, { "epoch": 3.996501901140684, "grad_norm": 0.4375, "learning_rate": 1.4860601157962173e-09, "loss": 1.6436, "num_input_tokens_seen": 861020160, "step": 6570, "train_runtime": 270344.7392, "train_tokens_per_second": 3184.897 }, { "epoch": 4.0, "num_input_tokens_seen": 861765632, "step": 6576, "total_flos": 1.457459776629139e+19, "train_loss": 2.710902440866995, "train_runtime": 270642.3692, "train_samples_per_second": 1.555, "train_steps_per_second": 0.024, "train_tokens_per_second": 3184.149 } ], "logging_steps": 10, "max_steps": 6576, "num_input_tokens_seen": 861765632, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.457459776629139e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }