StepCount-7b-SFT-16k / trainer_state.json
MING-ZCH's picture
Upload folder using huggingface_hub
66630be verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 3099,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0048408568316592035,
"grad_norm": 178.71523526371087,
"learning_rate": 6.451612903225807e-07,
"loss": 15.8295,
"num_input_tokens_seen": 1971904,
"step": 5,
"train_runtime": 965.9569,
"train_tokens_per_second": 2041.4
},
{
"epoch": 0.009681713663318407,
"grad_norm": 157.57466219746485,
"learning_rate": 1.4516129032258066e-06,
"loss": 15.3582,
"num_input_tokens_seen": 4038576,
"step": 10,
"train_runtime": 1957.3451,
"train_tokens_per_second": 2063.293
},
{
"epoch": 0.014522570494977611,
"grad_norm": 156.50950938238927,
"learning_rate": 2.2580645161290324e-06,
"loss": 12.8373,
"num_input_tokens_seen": 5891616,
"step": 15,
"train_runtime": 2767.3168,
"train_tokens_per_second": 2128.999
},
{
"epoch": 0.019363427326636814,
"grad_norm": 114.22743265072562,
"learning_rate": 3.064516129032258e-06,
"loss": 9.4205,
"num_input_tokens_seen": 7849472,
"step": 20,
"train_runtime": 3522.7348,
"train_tokens_per_second": 2228.232
},
{
"epoch": 0.024204284158296018,
"grad_norm": 135.40188397844682,
"learning_rate": 3.870967741935484e-06,
"loss": 7.6571,
"num_input_tokens_seen": 9689648,
"step": 25,
"train_runtime": 3820.9748,
"train_tokens_per_second": 2535.91
},
{
"epoch": 0.029045140989955223,
"grad_norm": 210.55574450271726,
"learning_rate": 4.67741935483871e-06,
"loss": 6.7016,
"num_input_tokens_seen": 11652144,
"step": 30,
"train_runtime": 4196.5025,
"train_tokens_per_second": 2776.632
},
{
"epoch": 0.03388599782161442,
"grad_norm": 208.41807492778292,
"learning_rate": 5.483870967741936e-06,
"loss": 5.8834,
"num_input_tokens_seen": 13544176,
"step": 35,
"train_runtime": 4483.8305,
"train_tokens_per_second": 3020.671
},
{
"epoch": 0.03872685465327363,
"grad_norm": 202.64781308250403,
"learning_rate": 6.290322580645161e-06,
"loss": 5.0909,
"num_input_tokens_seen": 15315344,
"step": 40,
"train_runtime": 4739.2398,
"train_tokens_per_second": 3231.604
},
{
"epoch": 0.04356771148493283,
"grad_norm": 208.02120960809935,
"learning_rate": 7.096774193548387e-06,
"loss": 4.3397,
"num_input_tokens_seen": 17094640,
"step": 45,
"train_runtime": 4992.3854,
"train_tokens_per_second": 3424.143
},
{
"epoch": 0.048408568316592036,
"grad_norm": 200.84606442004377,
"learning_rate": 7.903225806451613e-06,
"loss": 3.5344,
"num_input_tokens_seen": 18990000,
"step": 50,
"train_runtime": 5321.5443,
"train_tokens_per_second": 3568.513
},
{
"epoch": 0.05324942514825124,
"grad_norm": 180.4972165263306,
"learning_rate": 8.70967741935484e-06,
"loss": 2.5445,
"num_input_tokens_seen": 20854912,
"step": 55,
"train_runtime": 5860.7159,
"train_tokens_per_second": 3558.424
},
{
"epoch": 0.058090281979910445,
"grad_norm": 142.268336355056,
"learning_rate": 9.516129032258064e-06,
"loss": 1.5114,
"num_input_tokens_seen": 22734480,
"step": 60,
"train_runtime": 6689.9051,
"train_tokens_per_second": 3398.326
},
{
"epoch": 0.06293113881156964,
"grad_norm": 57.87038800361589,
"learning_rate": 1.0322580645161291e-05,
"loss": 0.681,
"num_input_tokens_seen": 24705984,
"step": 65,
"train_runtime": 7695.5578,
"train_tokens_per_second": 3210.421
},
{
"epoch": 0.06777199564322885,
"grad_norm": 11.382633199665975,
"learning_rate": 1.1129032258064517e-05,
"loss": 0.3083,
"num_input_tokens_seen": 26623312,
"step": 70,
"train_runtime": 8523.5587,
"train_tokens_per_second": 3123.497
},
{
"epoch": 0.07261285247488805,
"grad_norm": 1.7740955778794014,
"learning_rate": 1.1935483870967743e-05,
"loss": 0.2031,
"num_input_tokens_seen": 28560464,
"step": 75,
"train_runtime": 9449.2243,
"train_tokens_per_second": 3022.519
},
{
"epoch": 0.07745370930654726,
"grad_norm": 2.7185275952966643,
"learning_rate": 1.2741935483870968e-05,
"loss": 0.1776,
"num_input_tokens_seen": 30353664,
"step": 80,
"train_runtime": 10242.2399,
"train_tokens_per_second": 2963.577
},
{
"epoch": 0.08229456613820646,
"grad_norm": 11.173889555440166,
"learning_rate": 1.3548387096774195e-05,
"loss": 0.1556,
"num_input_tokens_seen": 32276544,
"step": 85,
"train_runtime": 11127.6455,
"train_tokens_per_second": 2900.573
},
{
"epoch": 0.08713542296986566,
"grad_norm": 0.43982142593021933,
"learning_rate": 1.4354838709677421e-05,
"loss": 0.1393,
"num_input_tokens_seen": 34137648,
"step": 90,
"train_runtime": 11810.6178,
"train_tokens_per_second": 2890.42
},
{
"epoch": 0.09197627980152487,
"grad_norm": 0.41292625320897214,
"learning_rate": 1.5161290322580646e-05,
"loss": 0.1302,
"num_input_tokens_seen": 35998352,
"step": 95,
"train_runtime": 12094.7097,
"train_tokens_per_second": 2976.372
},
{
"epoch": 0.09681713663318407,
"grad_norm": 0.2858556913615977,
"learning_rate": 1.596774193548387e-05,
"loss": 0.1201,
"num_input_tokens_seen": 37869184,
"step": 100,
"train_runtime": 12394.8839,
"train_tokens_per_second": 3055.227
},
{
"epoch": 0.10165799346484328,
"grad_norm": 0.30648690802872136,
"learning_rate": 1.6774193548387098e-05,
"loss": 0.1209,
"num_input_tokens_seen": 39694272,
"step": 105,
"train_runtime": 12703.266,
"train_tokens_per_second": 3124.73
},
{
"epoch": 0.10649885029650248,
"grad_norm": 0.2626276819380433,
"learning_rate": 1.7580645161290325e-05,
"loss": 0.1209,
"num_input_tokens_seen": 41530000,
"step": 110,
"train_runtime": 12989.8073,
"train_tokens_per_second": 3197.122
},
{
"epoch": 0.11133970712816169,
"grad_norm": 0.25596054422612646,
"learning_rate": 1.838709677419355e-05,
"loss": 0.1105,
"num_input_tokens_seen": 43396048,
"step": 115,
"train_runtime": 13306.8176,
"train_tokens_per_second": 3261.189
},
{
"epoch": 0.11618056395982089,
"grad_norm": 0.233381326668112,
"learning_rate": 1.9193548387096774e-05,
"loss": 0.1074,
"num_input_tokens_seen": 45343968,
"step": 120,
"train_runtime": 14110.9269,
"train_tokens_per_second": 3213.394
},
{
"epoch": 0.1210214207914801,
"grad_norm": 0.21251277353342254,
"learning_rate": 2e-05,
"loss": 0.1037,
"num_input_tokens_seen": 47322080,
"step": 125,
"train_runtime": 15041.1565,
"train_tokens_per_second": 3146.173
},
{
"epoch": 0.12586227762313928,
"grad_norm": 0.2127294082203629,
"learning_rate": 2.0806451612903226e-05,
"loss": 0.1062,
"num_input_tokens_seen": 49256592,
"step": 130,
"train_runtime": 16039.8532,
"train_tokens_per_second": 3070.888
},
{
"epoch": 0.1307031344547985,
"grad_norm": 0.37877204299340766,
"learning_rate": 2.1612903225806454e-05,
"loss": 0.1022,
"num_input_tokens_seen": 51154720,
"step": 135,
"train_runtime": 16882.3542,
"train_tokens_per_second": 3030.07
},
{
"epoch": 0.1355439912864577,
"grad_norm": 0.23637714784511715,
"learning_rate": 2.2419354838709678e-05,
"loss": 0.0953,
"num_input_tokens_seen": 53264704,
"step": 140,
"train_runtime": 18022.3173,
"train_tokens_per_second": 2955.486
},
{
"epoch": 0.1403848481181169,
"grad_norm": 0.25592037593094186,
"learning_rate": 2.3225806451612906e-05,
"loss": 0.1071,
"num_input_tokens_seen": 55099456,
"step": 145,
"train_runtime": 18802.202,
"train_tokens_per_second": 2930.479
},
{
"epoch": 0.1452257049497761,
"grad_norm": 0.24437969699285353,
"learning_rate": 2.403225806451613e-05,
"loss": 0.1,
"num_input_tokens_seen": 57033824,
"step": 150,
"train_runtime": 19705.9937,
"train_tokens_per_second": 2894.237
},
{
"epoch": 0.15006656178143532,
"grad_norm": 0.21535509944605138,
"learning_rate": 2.4838709677419354e-05,
"loss": 0.1024,
"num_input_tokens_seen": 58876656,
"step": 155,
"train_runtime": 20511.3799,
"train_tokens_per_second": 2870.439
},
{
"epoch": 0.1549074186130945,
"grad_norm": 0.2173650300325744,
"learning_rate": 2.5645161290322582e-05,
"loss": 0.0978,
"num_input_tokens_seen": 60877568,
"step": 160,
"train_runtime": 21475.5689,
"train_tokens_per_second": 2834.736
},
{
"epoch": 0.15974827544475373,
"grad_norm": 0.22756324955154553,
"learning_rate": 2.645161290322581e-05,
"loss": 0.1081,
"num_input_tokens_seen": 62756608,
"step": 165,
"train_runtime": 22319.9154,
"train_tokens_per_second": 2811.687
},
{
"epoch": 0.16458913227641292,
"grad_norm": 0.17479824979429293,
"learning_rate": 2.7258064516129034e-05,
"loss": 0.0934,
"num_input_tokens_seen": 64688544,
"step": 170,
"train_runtime": 23223.6296,
"train_tokens_per_second": 2785.462
},
{
"epoch": 0.16942998910807214,
"grad_norm": 0.18928070472014152,
"learning_rate": 2.806451612903226e-05,
"loss": 0.0956,
"num_input_tokens_seen": 66503488,
"step": 175,
"train_runtime": 23940.3182,
"train_tokens_per_second": 2777.887
},
{
"epoch": 0.17427084593973133,
"grad_norm": 0.19529886904869875,
"learning_rate": 2.8870967741935483e-05,
"loss": 0.0945,
"num_input_tokens_seen": 68426064,
"step": 180,
"train_runtime": 24825.8993,
"train_tokens_per_second": 2756.237
},
{
"epoch": 0.17911170277139055,
"grad_norm": 0.19453179688267283,
"learning_rate": 2.967741935483871e-05,
"loss": 0.0957,
"num_input_tokens_seen": 70300224,
"step": 185,
"train_runtime": 25652.738,
"train_tokens_per_second": 2740.457
},
{
"epoch": 0.18395255960304974,
"grad_norm": 0.21743752742438213,
"learning_rate": 3.0483870967741935e-05,
"loss": 0.1048,
"num_input_tokens_seen": 72153584,
"step": 190,
"train_runtime": 26532.9803,
"train_tokens_per_second": 2719.392
},
{
"epoch": 0.18879341643470895,
"grad_norm": 0.17245846000709703,
"learning_rate": 3.1290322580645166e-05,
"loss": 0.1038,
"num_input_tokens_seen": 73989040,
"step": 195,
"train_runtime": 27354.3265,
"train_tokens_per_second": 2704.839
},
{
"epoch": 0.19363427326636815,
"grad_norm": 0.18251260903855057,
"learning_rate": 3.2096774193548393e-05,
"loss": 0.106,
"num_input_tokens_seen": 75844448,
"step": 200,
"train_runtime": 28180.8862,
"train_tokens_per_second": 2691.344
},
{
"epoch": 0.19847513009802736,
"grad_norm": 0.17893448735702916,
"learning_rate": 3.2903225806451614e-05,
"loss": 0.0979,
"num_input_tokens_seen": 77627520,
"step": 205,
"train_runtime": 28955.3576,
"train_tokens_per_second": 2680.938
},
{
"epoch": 0.20331598692968655,
"grad_norm": 0.20201575024600307,
"learning_rate": 3.370967741935484e-05,
"loss": 0.0953,
"num_input_tokens_seen": 79416736,
"step": 210,
"train_runtime": 29709.3856,
"train_tokens_per_second": 2673.119
},
{
"epoch": 0.20815684376134574,
"grad_norm": 0.1752100891136059,
"learning_rate": 3.451612903225806e-05,
"loss": 0.0936,
"num_input_tokens_seen": 81375920,
"step": 215,
"train_runtime": 30576.7229,
"train_tokens_per_second": 2661.368
},
{
"epoch": 0.21299770059300496,
"grad_norm": 0.1691448592145883,
"learning_rate": 3.532258064516129e-05,
"loss": 0.0954,
"num_input_tokens_seen": 83230304,
"step": 220,
"train_runtime": 31349.3154,
"train_tokens_per_second": 2654.932
},
{
"epoch": 0.21783855742466415,
"grad_norm": 0.19940515973456982,
"learning_rate": 3.612903225806452e-05,
"loss": 0.095,
"num_input_tokens_seen": 85271952,
"step": 225,
"train_runtime": 32261.937,
"train_tokens_per_second": 2643.113
},
{
"epoch": 0.22267941425632337,
"grad_norm": 0.20741829321407812,
"learning_rate": 3.6935483870967746e-05,
"loss": 0.0898,
"num_input_tokens_seen": 87088336,
"step": 230,
"train_runtime": 33023.2573,
"train_tokens_per_second": 2637.182
},
{
"epoch": 0.22752027108798256,
"grad_norm": 0.19208547326415293,
"learning_rate": 3.7741935483870974e-05,
"loss": 0.1018,
"num_input_tokens_seen": 88804864,
"step": 235,
"train_runtime": 33715.7377,
"train_tokens_per_second": 2633.929
},
{
"epoch": 0.23236112791964178,
"grad_norm": 0.1778620676104918,
"learning_rate": 3.8548387096774195e-05,
"loss": 0.1003,
"num_input_tokens_seen": 90678512,
"step": 240,
"train_runtime": 34461.3908,
"train_tokens_per_second": 2631.307
},
{
"epoch": 0.23720198475130097,
"grad_norm": 0.19195813050583843,
"learning_rate": 3.935483870967742e-05,
"loss": 0.1012,
"num_input_tokens_seen": 92516880,
"step": 245,
"train_runtime": 35260.5603,
"train_tokens_per_second": 2623.806
},
{
"epoch": 0.2420428415829602,
"grad_norm": 0.1935559412076288,
"learning_rate": 4.016129032258065e-05,
"loss": 0.0944,
"num_input_tokens_seen": 94387008,
"step": 250,
"train_runtime": 36080.7203,
"train_tokens_per_second": 2615.996
},
{
"epoch": 0.24688369841461938,
"grad_norm": 0.1871264431371236,
"learning_rate": 4.096774193548387e-05,
"loss": 0.0952,
"num_input_tokens_seen": 96291040,
"step": 255,
"train_runtime": 36929.8595,
"train_tokens_per_second": 2607.403
},
{
"epoch": 0.25172455524627857,
"grad_norm": 0.16780208844120814,
"learning_rate": 4.17741935483871e-05,
"loss": 0.1029,
"num_input_tokens_seen": 98122448,
"step": 260,
"train_runtime": 37715.1617,
"train_tokens_per_second": 2601.671
},
{
"epoch": 0.2565654120779378,
"grad_norm": 0.15596752998410582,
"learning_rate": 4.258064516129032e-05,
"loss": 0.0955,
"num_input_tokens_seen": 100020384,
"step": 265,
"train_runtime": 38487.5175,
"train_tokens_per_second": 2598.775
},
{
"epoch": 0.261406268909597,
"grad_norm": 0.1887794985586671,
"learning_rate": 4.3387096774193554e-05,
"loss": 0.0966,
"num_input_tokens_seen": 101811184,
"step": 270,
"train_runtime": 39201.4963,
"train_tokens_per_second": 2597.125
},
{
"epoch": 0.2662471257412562,
"grad_norm": 0.21763986935352733,
"learning_rate": 4.4193548387096775e-05,
"loss": 0.0977,
"num_input_tokens_seen": 103544960,
"step": 275,
"train_runtime": 39839.3963,
"train_tokens_per_second": 2599.059
},
{
"epoch": 0.2710879825729154,
"grad_norm": 0.1422251923467383,
"learning_rate": 4.5e-05,
"loss": 0.0941,
"num_input_tokens_seen": 105431200,
"step": 280,
"train_runtime": 40502.3645,
"train_tokens_per_second": 2603.088
},
{
"epoch": 0.27592883940457463,
"grad_norm": 0.19060031607841202,
"learning_rate": 4.580645161290323e-05,
"loss": 0.0973,
"num_input_tokens_seen": 107368704,
"step": 285,
"train_runtime": 40864.2265,
"train_tokens_per_second": 2627.45
},
{
"epoch": 0.2807696962362338,
"grad_norm": 0.15547346024027353,
"learning_rate": 4.661290322580645e-05,
"loss": 0.0949,
"num_input_tokens_seen": 109257200,
"step": 290,
"train_runtime": 41191.3176,
"train_tokens_per_second": 2652.433
},
{
"epoch": 0.285610553067893,
"grad_norm": 0.18137966749300305,
"learning_rate": 4.741935483870968e-05,
"loss": 0.0953,
"num_input_tokens_seen": 111150704,
"step": 295,
"train_runtime": 41496.46,
"train_tokens_per_second": 2678.559
},
{
"epoch": 0.2904514098995522,
"grad_norm": 0.1726911372012115,
"learning_rate": 4.822580645161291e-05,
"loss": 0.097,
"num_input_tokens_seen": 113034512,
"step": 300,
"train_runtime": 41825.6854,
"train_tokens_per_second": 2702.514
},
{
"epoch": 0.29529226673121145,
"grad_norm": 0.14521468889499597,
"learning_rate": 4.903225806451613e-05,
"loss": 0.0897,
"num_input_tokens_seen": 114937056,
"step": 305,
"train_runtime": 42162.0241,
"train_tokens_per_second": 2726.08
},
{
"epoch": 0.30013312356287064,
"grad_norm": 0.15231521973585768,
"learning_rate": 4.983870967741936e-05,
"loss": 0.0887,
"num_input_tokens_seen": 116955552,
"step": 310,
"train_runtime": 42539.7887,
"train_tokens_per_second": 2749.321
},
{
"epoch": 0.30497398039452983,
"grad_norm": 0.1534116557864518,
"learning_rate": 4.999974623485097e-05,
"loss": 0.0981,
"num_input_tokens_seen": 118830688,
"step": 315,
"train_runtime": 42871.2685,
"train_tokens_per_second": 2771.802
},
{
"epoch": 0.309814837226189,
"grad_norm": 0.1686940018305674,
"learning_rate": 4.999871532276243e-05,
"loss": 0.0974,
"num_input_tokens_seen": 120691504,
"step": 320,
"train_runtime": 43165.3478,
"train_tokens_per_second": 2796.028
},
{
"epoch": 0.3146556940578482,
"grad_norm": 0.16433905364856274,
"learning_rate": 4.999689143608857e-05,
"loss": 0.0948,
"num_input_tokens_seen": 122560384,
"step": 325,
"train_runtime": 43458.1159,
"train_tokens_per_second": 2820.196
},
{
"epoch": 0.31949655088950746,
"grad_norm": 0.14174604664203966,
"learning_rate": 4.999427463268419e-05,
"loss": 0.0901,
"num_input_tokens_seen": 124493200,
"step": 330,
"train_runtime": 43797.1282,
"train_tokens_per_second": 2842.497
},
{
"epoch": 0.32433740772116665,
"grad_norm": 0.1695048202968491,
"learning_rate": 4.999086499555591e-05,
"loss": 0.09,
"num_input_tokens_seen": 126353344,
"step": 335,
"train_runtime": 44105.2533,
"train_tokens_per_second": 2864.814
},
{
"epoch": 0.32917826455282584,
"grad_norm": 0.17629027656795873,
"learning_rate": 4.9986662632859504e-05,
"loss": 0.0943,
"num_input_tokens_seen": 128177040,
"step": 340,
"train_runtime": 44411.4535,
"train_tokens_per_second": 2886.126
},
{
"epoch": 0.33401912138448503,
"grad_norm": 0.16367775596487932,
"learning_rate": 4.9981667677896506e-05,
"loss": 0.0921,
"num_input_tokens_seen": 129970848,
"step": 345,
"train_runtime": 44670.9184,
"train_tokens_per_second": 2909.518
},
{
"epoch": 0.3388599782161443,
"grad_norm": 0.14429236107461654,
"learning_rate": 4.997588028910993e-05,
"loss": 0.0933,
"num_input_tokens_seen": 131934512,
"step": 350,
"train_runtime": 45023.4447,
"train_tokens_per_second": 2930.351
},
{
"epoch": 0.34370083504780347,
"grad_norm": 0.14127408264764416,
"learning_rate": 4.996930065007932e-05,
"loss": 0.0919,
"num_input_tokens_seen": 133828000,
"step": 355,
"train_runtime": 45336.3241,
"train_tokens_per_second": 2951.893
},
{
"epoch": 0.34854169187946266,
"grad_norm": 0.1480021003298484,
"learning_rate": 4.996192896951485e-05,
"loss": 0.0894,
"num_input_tokens_seen": 135655808,
"step": 360,
"train_runtime": 45602.4522,
"train_tokens_per_second": 2974.748
},
{
"epoch": 0.35338254871112185,
"grad_norm": 0.16542232457655492,
"learning_rate": 4.995376548125075e-05,
"loss": 0.089,
"num_input_tokens_seen": 137547952,
"step": 365,
"train_runtime": 45922.7124,
"train_tokens_per_second": 2995.205
},
{
"epoch": 0.3582234055427811,
"grad_norm": 0.15136452715359272,
"learning_rate": 4.994481044423788e-05,
"loss": 0.0963,
"num_input_tokens_seen": 139408624,
"step": 370,
"train_runtime": 46232.1379,
"train_tokens_per_second": 3015.405
},
{
"epoch": 0.3630642623744403,
"grad_norm": 0.1569038436076336,
"learning_rate": 4.99350641425355e-05,
"loss": 0.098,
"num_input_tokens_seen": 141146720,
"step": 375,
"train_runtime": 46495.1626,
"train_tokens_per_second": 3035.729
},
{
"epoch": 0.3679051192060995,
"grad_norm": 0.15583043362247376,
"learning_rate": 4.99245268853023e-05,
"loss": 0.0927,
"num_input_tokens_seen": 142875840,
"step": 380,
"train_runtime": 46725.0644,
"train_tokens_per_second": 3057.799
},
{
"epoch": 0.37274597603775866,
"grad_norm": 0.17120656256693706,
"learning_rate": 4.9913199006786515e-05,
"loss": 0.0881,
"num_input_tokens_seen": 144878720,
"step": 385,
"train_runtime": 47082.6128,
"train_tokens_per_second": 3077.117
},
{
"epoch": 0.3775868328694179,
"grad_norm": 0.16312830261625755,
"learning_rate": 4.9901080866315416e-05,
"loss": 0.1036,
"num_input_tokens_seen": 146604080,
"step": 390,
"train_runtime": 47321.9205,
"train_tokens_per_second": 3098.016
},
{
"epoch": 0.3824276897010771,
"grad_norm": 0.11584886103538528,
"learning_rate": 4.988817284828386e-05,
"loss": 0.0879,
"num_input_tokens_seen": 148505760,
"step": 395,
"train_runtime": 47625.9364,
"train_tokens_per_second": 3118.17
},
{
"epoch": 0.3872685465327363,
"grad_norm": 0.1146886026314572,
"learning_rate": 4.9874475362142076e-05,
"loss": 0.0893,
"num_input_tokens_seen": 150426000,
"step": 400,
"train_runtime": 47956.2272,
"train_tokens_per_second": 3136.735
},
{
"epoch": 0.3921094033643955,
"grad_norm": 0.12910217232559762,
"learning_rate": 4.9859988842382744e-05,
"loss": 0.0909,
"num_input_tokens_seen": 152243872,
"step": 405,
"train_runtime": 48242.6162,
"train_tokens_per_second": 3155.796
},
{
"epoch": 0.3969502601960547,
"grad_norm": 0.15108152268102795,
"learning_rate": 4.984471374852715e-05,
"loss": 0.0947,
"num_input_tokens_seen": 154052240,
"step": 410,
"train_runtime": 48527.6459,
"train_tokens_per_second": 3174.525
},
{
"epoch": 0.4017911170277139,
"grad_norm": 0.14679850444524517,
"learning_rate": 4.9828650565110644e-05,
"loss": 0.0907,
"num_input_tokens_seen": 156020992,
"step": 415,
"train_runtime": 48888.6576,
"train_tokens_per_second": 3191.354
},
{
"epoch": 0.4066319738593731,
"grad_norm": 0.1327811810612439,
"learning_rate": 4.981179980166726e-05,
"loss": 0.09,
"num_input_tokens_seen": 157936240,
"step": 420,
"train_runtime": 49230.7981,
"train_tokens_per_second": 3208.078
},
{
"epoch": 0.4114728306910323,
"grad_norm": 0.14604735675352598,
"learning_rate": 4.979416199271354e-05,
"loss": 0.0903,
"num_input_tokens_seen": 159733040,
"step": 425,
"train_runtime": 49491.6444,
"train_tokens_per_second": 3227.475
},
{
"epoch": 0.4163136875226915,
"grad_norm": 0.14435246296113152,
"learning_rate": 4.977573769773162e-05,
"loss": 0.093,
"num_input_tokens_seen": 161662848,
"step": 430,
"train_runtime": 49847.5614,
"train_tokens_per_second": 3243.145
},
{
"epoch": 0.42115454435435073,
"grad_norm": 0.14601625994718231,
"learning_rate": 4.975652750115143e-05,
"loss": 0.0953,
"num_input_tokens_seen": 163510512,
"step": 435,
"train_runtime": 50144.9198,
"train_tokens_per_second": 3260.759
},
{
"epoch": 0.4259954011860099,
"grad_norm": 0.13966377448924783,
"learning_rate": 4.973653201233219e-05,
"loss": 0.0925,
"num_input_tokens_seen": 165482800,
"step": 440,
"train_runtime": 50546.5604,
"train_tokens_per_second": 3273.869
},
{
"epoch": 0.4308362580176691,
"grad_norm": 0.1417296980128132,
"learning_rate": 4.971575186554307e-05,
"loss": 0.0893,
"num_input_tokens_seen": 167426896,
"step": 445,
"train_runtime": 50915.2604,
"train_tokens_per_second": 3288.344
},
{
"epoch": 0.4356771148493283,
"grad_norm": 0.13565426401676745,
"learning_rate": 4.969418771994309e-05,
"loss": 0.0943,
"num_input_tokens_seen": 169233824,
"step": 450,
"train_runtime": 51185.0637,
"train_tokens_per_second": 3306.313
},
{
"epoch": 0.44051797168098755,
"grad_norm": 0.1454259863391591,
"learning_rate": 4.967184025956015e-05,
"loss": 0.0938,
"num_input_tokens_seen": 171055232,
"step": 455,
"train_runtime": 51458.1503,
"train_tokens_per_second": 3324.162
},
{
"epoch": 0.44535882851264674,
"grad_norm": 0.13730748863216408,
"learning_rate": 4.9648710193269424e-05,
"loss": 0.0915,
"num_input_tokens_seen": 172893920,
"step": 460,
"train_runtime": 51744.3108,
"train_tokens_per_second": 3341.313
},
{
"epoch": 0.45019968534430593,
"grad_norm": 0.12243741977778125,
"learning_rate": 4.962479825477079e-05,
"loss": 0.0849,
"num_input_tokens_seen": 174782016,
"step": 465,
"train_runtime": 52060.903,
"train_tokens_per_second": 3357.261
},
{
"epoch": 0.4550405421759651,
"grad_norm": 0.11854640411216295,
"learning_rate": 4.9600105202565605e-05,
"loss": 0.0886,
"num_input_tokens_seen": 176671136,
"step": 470,
"train_runtime": 52381.9286,
"train_tokens_per_second": 3372.75
},
{
"epoch": 0.45988139900762437,
"grad_norm": 0.14968248763796943,
"learning_rate": 4.9574631819932646e-05,
"loss": 0.0924,
"num_input_tokens_seen": 178524848,
"step": 475,
"train_runtime": 52697.5978,
"train_tokens_per_second": 3387.723
},
{
"epoch": 0.46472225583928356,
"grad_norm": 0.1435093288248635,
"learning_rate": 4.9548378914903225e-05,
"loss": 0.0872,
"num_input_tokens_seen": 180401664,
"step": 480,
"train_runtime": 53014.3662,
"train_tokens_per_second": 3402.883
},
{
"epoch": 0.46956311267094275,
"grad_norm": 0.15828157632384718,
"learning_rate": 4.95213473202356e-05,
"loss": 0.0948,
"num_input_tokens_seen": 182234608,
"step": 485,
"train_runtime": 53294.1919,
"train_tokens_per_second": 3419.408
},
{
"epoch": 0.47440396950260194,
"grad_norm": 0.11560281909563656,
"learning_rate": 4.9493537893388544e-05,
"loss": 0.0884,
"num_input_tokens_seen": 184157296,
"step": 490,
"train_runtime": 53645.7615,
"train_tokens_per_second": 3432.84
},
{
"epoch": 0.4792448263342612,
"grad_norm": 0.14531956629532314,
"learning_rate": 4.9464951516494116e-05,
"loss": 0.0916,
"num_input_tokens_seen": 185978048,
"step": 495,
"train_runtime": 53924.698,
"train_tokens_per_second": 3448.847
},
{
"epoch": 0.4840856831659204,
"grad_norm": 0.11828670163218845,
"learning_rate": 4.943558909632975e-05,
"loss": 0.0877,
"num_input_tokens_seen": 187867312,
"step": 500,
"train_runtime": 54232.0578,
"train_tokens_per_second": 3464.138
},
{
"epoch": 0.48892653999757957,
"grad_norm": 0.1315673057494712,
"learning_rate": 4.94054515642894e-05,
"loss": 0.0962,
"num_input_tokens_seen": 189760400,
"step": 505,
"train_runtime": 54567.8241,
"train_tokens_per_second": 3477.515
},
{
"epoch": 0.49376739682923876,
"grad_norm": 0.13676145087673666,
"learning_rate": 4.937453987635408e-05,
"loss": 0.084,
"num_input_tokens_seen": 191773264,
"step": 510,
"train_runtime": 54988.129,
"train_tokens_per_second": 3487.539
},
{
"epoch": 0.498608253660898,
"grad_norm": 0.1274509635019909,
"learning_rate": 4.934285501306152e-05,
"loss": 0.0868,
"num_input_tokens_seen": 193750176,
"step": 515,
"train_runtime": 55385.4065,
"train_tokens_per_second": 3498.217
},
{
"epoch": 0.5034491104925571,
"grad_norm": 0.12982283309243445,
"learning_rate": 4.931039797947498e-05,
"loss": 0.0895,
"num_input_tokens_seen": 195664256,
"step": 520,
"train_runtime": 55687.4295,
"train_tokens_per_second": 3513.616
},
{
"epoch": 0.5082899673242164,
"grad_norm": 0.13281239797421768,
"learning_rate": 4.9277169805151496e-05,
"loss": 0.0875,
"num_input_tokens_seen": 197577536,
"step": 525,
"train_runtime": 56037.147,
"train_tokens_per_second": 3525.831
},
{
"epoch": 0.5131308241558756,
"grad_norm": 0.127865250498207,
"learning_rate": 4.924317154410915e-05,
"loss": 0.0907,
"num_input_tokens_seen": 199441664,
"step": 530,
"train_runtime": 56386.1137,
"train_tokens_per_second": 3537.071
},
{
"epoch": 0.5179716809875348,
"grad_norm": 0.12535903140665275,
"learning_rate": 4.9208404274793615e-05,
"loss": 0.0851,
"num_input_tokens_seen": 201395760,
"step": 535,
"train_runtime": 56725.1327,
"train_tokens_per_second": 3550.38
},
{
"epoch": 0.522812537819194,
"grad_norm": 0.11973273742173074,
"learning_rate": 4.917286910004402e-05,
"loss": 0.0859,
"num_input_tokens_seen": 203289168,
"step": 540,
"train_runtime": 57048.2676,
"train_tokens_per_second": 3563.459
},
{
"epoch": 0.5276533946508531,
"grad_norm": 0.13617614252406723,
"learning_rate": 4.9136567147057886e-05,
"loss": 0.0932,
"num_input_tokens_seen": 205086960,
"step": 545,
"train_runtime": 57320.4179,
"train_tokens_per_second": 3577.904
},
{
"epoch": 0.5324942514825124,
"grad_norm": 0.12228835012992155,
"learning_rate": 4.9099499567355465e-05,
"loss": 0.0877,
"num_input_tokens_seen": 207038560,
"step": 550,
"train_runtime": 57676.6952,
"train_tokens_per_second": 3589.64
},
{
"epoch": 0.5373351083141716,
"grad_norm": 0.11616581415998756,
"learning_rate": 4.9061667536743096e-05,
"loss": 0.0849,
"num_input_tokens_seen": 208994656,
"step": 555,
"train_runtime": 58017.1366,
"train_tokens_per_second": 3602.292
},
{
"epoch": 0.5421759651458308,
"grad_norm": 0.1227132194885445,
"learning_rate": 4.9023072255276e-05,
"loss": 0.0924,
"num_input_tokens_seen": 210860592,
"step": 560,
"train_runtime": 58355.6811,
"train_tokens_per_second": 3613.369
},
{
"epoch": 0.54701682197749,
"grad_norm": 0.1331754681110147,
"learning_rate": 4.89837149472202e-05,
"loss": 0.0852,
"num_input_tokens_seen": 212822400,
"step": 565,
"train_runtime": 58727.1386,
"train_tokens_per_second": 3623.919
},
{
"epoch": 0.5518576788091493,
"grad_norm": 0.11555719706146601,
"learning_rate": 4.894359686101363e-05,
"loss": 0.0894,
"num_input_tokens_seen": 214750096,
"step": 570,
"train_runtime": 59082.0703,
"train_tokens_per_second": 3634.776
},
{
"epoch": 0.5566985356408084,
"grad_norm": 0.12553710069358454,
"learning_rate": 4.890271926922661e-05,
"loss": 0.0859,
"num_input_tokens_seen": 216744256,
"step": 575,
"train_runtime": 59467.6278,
"train_tokens_per_second": 3644.744
},
{
"epoch": 0.5615393924724676,
"grad_norm": 0.12296197236193043,
"learning_rate": 4.886108346852142e-05,
"loss": 0.0843,
"num_input_tokens_seen": 218618416,
"step": 580,
"train_runtime": 59760.4082,
"train_tokens_per_second": 3658.248
},
{
"epoch": 0.5663802493041268,
"grad_norm": 0.13282536365892234,
"learning_rate": 4.881869077961119e-05,
"loss": 0.0881,
"num_input_tokens_seen": 220473312,
"step": 585,
"train_runtime": 60072.1334,
"train_tokens_per_second": 3670.143
},
{
"epoch": 0.571221106135786,
"grad_norm": 0.1192404447399879,
"learning_rate": 4.8775542547218035e-05,
"loss": 0.0838,
"num_input_tokens_seen": 222434400,
"step": 590,
"train_runtime": 60432.6507,
"train_tokens_per_second": 3680.699
},
{
"epoch": 0.5760619629674453,
"grad_norm": 0.12137471825767925,
"learning_rate": 4.873164014003035e-05,
"loss": 0.0823,
"num_input_tokens_seen": 224344864,
"step": 595,
"train_runtime": 60761.8507,
"train_tokens_per_second": 3692.199
},
{
"epoch": 0.5809028197991044,
"grad_norm": 0.10630144681377657,
"learning_rate": 4.868698495065942e-05,
"loss": 0.0908,
"num_input_tokens_seen": 226325424,
"step": 600,
"train_runtime": 61173.0901,
"train_tokens_per_second": 3699.755
},
{
"epoch": 0.5857436766307637,
"grad_norm": 0.12903663659157497,
"learning_rate": 4.8641578395595244e-05,
"loss": 0.0841,
"num_input_tokens_seen": 228223680,
"step": 605,
"train_runtime": 61472.6009,
"train_tokens_per_second": 3712.608
},
{
"epoch": 0.5905845334624229,
"grad_norm": 0.1120445766713908,
"learning_rate": 4.8595421915161606e-05,
"loss": 0.09,
"num_input_tokens_seen": 230051616,
"step": 610,
"train_runtime": 61773.5624,
"train_tokens_per_second": 3724.111
},
{
"epoch": 0.595425390294082,
"grad_norm": 0.1403828596714754,
"learning_rate": 4.854851697347038e-05,
"loss": 0.0953,
"num_input_tokens_seen": 231784640,
"step": 615,
"train_runtime": 62017.5435,
"train_tokens_per_second": 3737.404
},
{
"epoch": 0.6002662471257413,
"grad_norm": 0.12026744008912688,
"learning_rate": 4.8500865058375084e-05,
"loss": 0.0818,
"num_input_tokens_seen": 233624880,
"step": 620,
"train_runtime": 62301.1537,
"train_tokens_per_second": 3749.929
},
{
"epoch": 0.6051071039574004,
"grad_norm": 0.13228231302304957,
"learning_rate": 4.845246768142371e-05,
"loss": 0.0919,
"num_input_tokens_seen": 235580064,
"step": 625,
"train_runtime": 62644.2313,
"train_tokens_per_second": 3760.603
},
{
"epoch": 0.6099479607890597,
"grad_norm": 0.11162603952629033,
"learning_rate": 4.840332637781072e-05,
"loss": 0.0806,
"num_input_tokens_seen": 237579456,
"step": 630,
"train_runtime": 63013.8952,
"train_tokens_per_second": 3770.271
},
{
"epoch": 0.6147888176207189,
"grad_norm": 0.13202350019993933,
"learning_rate": 4.835344270632844e-05,
"loss": 0.0894,
"num_input_tokens_seen": 239379872,
"step": 635,
"train_runtime": 63306.9343,
"train_tokens_per_second": 3781.258
},
{
"epoch": 0.619629674452378,
"grad_norm": 0.1262730134357675,
"learning_rate": 4.830281824931751e-05,
"loss": 0.0917,
"num_input_tokens_seen": 241343936,
"step": 640,
"train_runtime": 63660.7557,
"train_tokens_per_second": 3791.094
},
{
"epoch": 0.6244705312840373,
"grad_norm": 0.12514465647043127,
"learning_rate": 4.825145461261677e-05,
"loss": 0.0949,
"num_input_tokens_seen": 243126480,
"step": 645,
"train_runtime": 63956.927,
"train_tokens_per_second": 3801.41
},
{
"epoch": 0.6293113881156964,
"grad_norm": 0.11879787426434715,
"learning_rate": 4.8199353425512276e-05,
"loss": 0.0836,
"num_input_tokens_seen": 245036080,
"step": 650,
"train_runtime": 64282.9428,
"train_tokens_per_second": 3811.837
},
{
"epoch": 0.6341522449473557,
"grad_norm": 0.12807215290275195,
"learning_rate": 4.814651634068565e-05,
"loss": 0.0919,
"num_input_tokens_seen": 246836544,
"step": 655,
"train_runtime": 64554.9622,
"train_tokens_per_second": 3823.665
},
{
"epoch": 0.6389931017790149,
"grad_norm": 0.13318265083798486,
"learning_rate": 4.809294503416164e-05,
"loss": 0.0885,
"num_input_tokens_seen": 248715200,
"step": 660,
"train_runtime": 64862.1325,
"train_tokens_per_second": 3834.521
},
{
"epoch": 0.643833958610674,
"grad_norm": 0.11706198751336716,
"learning_rate": 4.8038641205254945e-05,
"loss": 0.0845,
"num_input_tokens_seen": 250584112,
"step": 665,
"train_runtime": 65162.2083,
"train_tokens_per_second": 3845.544
},
{
"epoch": 0.6486748154423333,
"grad_norm": 0.10849730614741822,
"learning_rate": 4.7983606576516335e-05,
"loss": 0.0834,
"num_input_tokens_seen": 252446000,
"step": 670,
"train_runtime": 65447.5256,
"train_tokens_per_second": 3857.228
},
{
"epoch": 0.6535156722739925,
"grad_norm": 0.11644681119826476,
"learning_rate": 4.792784289367799e-05,
"loss": 0.0889,
"num_input_tokens_seen": 254336512,
"step": 675,
"train_runtime": 65807.5992,
"train_tokens_per_second": 3864.85
},
{
"epoch": 0.6583565291056517,
"grad_norm": 0.12648798160840266,
"learning_rate": 4.787135192559814e-05,
"loss": 0.081,
"num_input_tokens_seen": 256248240,
"step": 680,
"train_runtime": 66118.3408,
"train_tokens_per_second": 3875.6
},
{
"epoch": 0.6631973859373109,
"grad_norm": 0.13179948151989662,
"learning_rate": 4.781413546420494e-05,
"loss": 0.0925,
"num_input_tokens_seen": 258018928,
"step": 685,
"train_runtime": 66397.3901,
"train_tokens_per_second": 3885.98
},
{
"epoch": 0.6680382427689701,
"grad_norm": 0.11355476378512105,
"learning_rate": 4.775619532443965e-05,
"loss": 0.081,
"num_input_tokens_seen": 260040928,
"step": 690,
"train_runtime": 66781.9966,
"train_tokens_per_second": 3893.878
},
{
"epoch": 0.6728790996006293,
"grad_norm": 0.1102920346549403,
"learning_rate": 4.7697533344199054e-05,
"loss": 0.081,
"num_input_tokens_seen": 261986240,
"step": 695,
"train_runtime": 67112.4268,
"train_tokens_per_second": 3903.692
},
{
"epoch": 0.6777199564322886,
"grad_norm": 0.11923784097224148,
"learning_rate": 4.763815138427713e-05,
"loss": 0.0834,
"num_input_tokens_seen": 263808880,
"step": 700,
"train_runtime": 67389.9808,
"train_tokens_per_second": 3914.66
},
{
"epoch": 0.6825608132639477,
"grad_norm": 0.13058866115819348,
"learning_rate": 4.757805132830609e-05,
"loss": 0.0858,
"num_input_tokens_seen": 265662032,
"step": 705,
"train_runtime": 67694.3239,
"train_tokens_per_second": 3924.436
},
{
"epoch": 0.6874016700956069,
"grad_norm": 0.10903443202471683,
"learning_rate": 4.751723508269659e-05,
"loss": 0.084,
"num_input_tokens_seen": 267501344,
"step": 710,
"train_runtime": 68025.3342,
"train_tokens_per_second": 3932.378
},
{
"epoch": 0.6922425269272662,
"grad_norm": 0.1065038057537401,
"learning_rate": 4.745570457657722e-05,
"loss": 0.0843,
"num_input_tokens_seen": 269394176,
"step": 715,
"train_runtime": 68343.6923,
"train_tokens_per_second": 3941.756
},
{
"epoch": 0.6970833837589253,
"grad_norm": 0.11228653074537234,
"learning_rate": 4.7393461761733395e-05,
"loss": 0.0911,
"num_input_tokens_seen": 271288208,
"step": 720,
"train_runtime": 68705.6867,
"train_tokens_per_second": 3948.555
},
{
"epoch": 0.7019242405905846,
"grad_norm": 0.10423741587634047,
"learning_rate": 4.733050861254538e-05,
"loss": 0.0759,
"num_input_tokens_seen": 273149952,
"step": 725,
"train_runtime": 69004.9126,
"train_tokens_per_second": 3958.413
},
{
"epoch": 0.7067650974222437,
"grad_norm": 0.12067320290718164,
"learning_rate": 4.7266847125925686e-05,
"loss": 0.0874,
"num_input_tokens_seen": 275017888,
"step": 730,
"train_runtime": 69310.7545,
"train_tokens_per_second": 3967.896
},
{
"epoch": 0.7116059542539029,
"grad_norm": 0.126476130474256,
"learning_rate": 4.720247932125572e-05,
"loss": 0.0877,
"num_input_tokens_seen": 276821728,
"step": 735,
"train_runtime": 69607.9909,
"train_tokens_per_second": 3976.867
},
{
"epoch": 0.7164468110855622,
"grad_norm": 0.14675215045590576,
"learning_rate": 4.713740724032173e-05,
"loss": 0.0854,
"num_input_tokens_seen": 278648688,
"step": 740,
"train_runtime": 69890.2265,
"train_tokens_per_second": 3986.948
},
{
"epoch": 0.7212876679172213,
"grad_norm": 0.10393986794688335,
"learning_rate": 4.7071632947250056e-05,
"loss": 0.0856,
"num_input_tokens_seen": 280573744,
"step": 745,
"train_runtime": 70220.8361,
"train_tokens_per_second": 3995.591
},
{
"epoch": 0.7261285247488806,
"grad_norm": 0.12245429208403327,
"learning_rate": 4.700515852844161e-05,
"loss": 0.088,
"num_input_tokens_seen": 282386080,
"step": 750,
"train_runtime": 70532.5051,
"train_tokens_per_second": 4003.63
},
{
"epoch": 0.7309693815805397,
"grad_norm": 0.15490079892082556,
"learning_rate": 4.6937986092505734e-05,
"loss": 0.0853,
"num_input_tokens_seen": 284163392,
"step": 755,
"train_runtime": 70777.4141,
"train_tokens_per_second": 4014.888
},
{
"epoch": 0.735810238412199,
"grad_norm": 0.1355824935484372,
"learning_rate": 4.687011777019331e-05,
"loss": 0.0897,
"num_input_tokens_seen": 285979472,
"step": 760,
"train_runtime": 71065.8783,
"train_tokens_per_second": 4024.146
},
{
"epoch": 0.7406510952438582,
"grad_norm": 0.10890370807480093,
"learning_rate": 4.680155571432915e-05,
"loss": 0.0866,
"num_input_tokens_seen": 287803904,
"step": 765,
"train_runtime": 71342.8364,
"train_tokens_per_second": 4034.097
},
{
"epoch": 0.7454919520755173,
"grad_norm": 0.10818940985882454,
"learning_rate": 4.673230209974372e-05,
"loss": 0.084,
"num_input_tokens_seen": 289770544,
"step": 770,
"train_runtime": 71692.4983,
"train_tokens_per_second": 4041.853
},
{
"epoch": 0.7503328089071766,
"grad_norm": 0.11393295664454559,
"learning_rate": 4.666235912320416e-05,
"loss": 0.0847,
"num_input_tokens_seen": 291639680,
"step": 775,
"train_runtime": 71991.759,
"train_tokens_per_second": 4051.015
},
{
"epoch": 0.7551736657388358,
"grad_norm": 0.10791548067564018,
"learning_rate": 4.6591729003344604e-05,
"loss": 0.0855,
"num_input_tokens_seen": 293493632,
"step": 780,
"train_runtime": 72290.0262,
"train_tokens_per_second": 4059.946
},
{
"epoch": 0.760014522570495,
"grad_norm": 0.12476206555821662,
"learning_rate": 4.652041398059577e-05,
"loss": 0.0874,
"num_input_tokens_seen": 295416672,
"step": 785,
"train_runtime": 72613.7692,
"train_tokens_per_second": 4068.329
},
{
"epoch": 0.7648553794021542,
"grad_norm": 0.11581041520648833,
"learning_rate": 4.644841631711393e-05,
"loss": 0.0917,
"num_input_tokens_seen": 297230848,
"step": 790,
"train_runtime": 72898.9964,
"train_tokens_per_second": 4077.297
},
{
"epoch": 0.7696962362338133,
"grad_norm": 0.1207889618919906,
"learning_rate": 4.637573829670913e-05,
"loss": 0.0862,
"num_input_tokens_seen": 299220432,
"step": 795,
"train_runtime": 73281.0367,
"train_tokens_per_second": 4083.19
},
{
"epoch": 0.7745370930654726,
"grad_norm": 0.13046576053931733,
"learning_rate": 4.630238222477277e-05,
"loss": 0.0876,
"num_input_tokens_seen": 301103984,
"step": 800,
"train_runtime": 73600.0892,
"train_tokens_per_second": 4091.082
},
{
"epoch": 0.7793779498971318,
"grad_norm": 0.11722404450994561,
"learning_rate": 4.622835042820445e-05,
"loss": 0.0832,
"num_input_tokens_seen": 303004240,
"step": 805,
"train_runtime": 73903.033,
"train_tokens_per_second": 4100.024
},
{
"epoch": 0.784218806728791,
"grad_norm": 0.11528092874791906,
"learning_rate": 4.615364525533817e-05,
"loss": 0.0875,
"num_input_tokens_seen": 304776752,
"step": 810,
"train_runtime": 74158.7012,
"train_tokens_per_second": 4109.791
},
{
"epoch": 0.7890596635604502,
"grad_norm": 0.1118105371174961,
"learning_rate": 4.6078269075867844e-05,
"loss": 0.0838,
"num_input_tokens_seen": 306564704,
"step": 815,
"train_runtime": 74444.0338,
"train_tokens_per_second": 4118.056
},
{
"epoch": 0.7939005203921095,
"grad_norm": 0.12405349561980368,
"learning_rate": 4.600222428077212e-05,
"loss": 0.0831,
"num_input_tokens_seen": 308443424,
"step": 820,
"train_runtime": 74777.6015,
"train_tokens_per_second": 4124.81
},
{
"epoch": 0.7987413772237686,
"grad_norm": 0.10166638613764127,
"learning_rate": 4.592551328223854e-05,
"loss": 0.0861,
"num_input_tokens_seen": 310326496,
"step": 825,
"train_runtime": 75075.0708,
"train_tokens_per_second": 4133.549
},
{
"epoch": 0.8035822340554278,
"grad_norm": 0.1304318839178686,
"learning_rate": 4.584813851358705e-05,
"loss": 0.0868,
"num_input_tokens_seen": 312324656,
"step": 830,
"train_runtime": 75496.3515,
"train_tokens_per_second": 4136.95
},
{
"epoch": 0.808423090887087,
"grad_norm": 0.11074066557269138,
"learning_rate": 4.577010242919277e-05,
"loss": 0.0823,
"num_input_tokens_seen": 314264240,
"step": 835,
"train_runtime": 75859.0005,
"train_tokens_per_second": 4142.742
},
{
"epoch": 0.8132639477187462,
"grad_norm": 0.1163799358603999,
"learning_rate": 4.569140750440817e-05,
"loss": 0.0823,
"num_input_tokens_seen": 316091424,
"step": 840,
"train_runtime": 76138.2624,
"train_tokens_per_second": 4151.545
},
{
"epoch": 0.8181048045504055,
"grad_norm": 0.11515834663049564,
"learning_rate": 4.561205623548453e-05,
"loss": 0.0816,
"num_input_tokens_seen": 318022304,
"step": 845,
"train_runtime": 76454.3449,
"train_tokens_per_second": 4159.637
},
{
"epoch": 0.8229456613820646,
"grad_norm": 0.10167337175748853,
"learning_rate": 4.5532051139492784e-05,
"loss": 0.0814,
"num_input_tokens_seen": 319889920,
"step": 850,
"train_runtime": 76782.5683,
"train_tokens_per_second": 4166.179
},
{
"epoch": 0.8277865182137238,
"grad_norm": 0.10009892012288468,
"learning_rate": 4.545139475424366e-05,
"loss": 0.0855,
"num_input_tokens_seen": 321805040,
"step": 855,
"train_runtime": 77138.4208,
"train_tokens_per_second": 4171.787
},
{
"epoch": 0.832627375045383,
"grad_norm": 0.10222131699800209,
"learning_rate": 4.537008963820717e-05,
"loss": 0.0859,
"num_input_tokens_seen": 323627168,
"step": 860,
"train_runtime": 77426.607,
"train_tokens_per_second": 4179.793
},
{
"epoch": 0.8374682318770422,
"grad_norm": 0.10643170835147328,
"learning_rate": 4.5288138370431464e-05,
"loss": 0.0866,
"num_input_tokens_seen": 325450784,
"step": 865,
"train_runtime": 77714.3766,
"train_tokens_per_second": 4187.781
},
{
"epoch": 0.8423090887087015,
"grad_norm": 0.1002014887379627,
"learning_rate": 4.520554355046105e-05,
"loss": 0.0794,
"num_input_tokens_seen": 327331824,
"step": 870,
"train_runtime": 78022.0251,
"train_tokens_per_second": 4195.377
},
{
"epoch": 0.8471499455403606,
"grad_norm": 0.12980276451699957,
"learning_rate": 4.512230779825427e-05,
"loss": 0.0853,
"num_input_tokens_seen": 329167072,
"step": 875,
"train_runtime": 78294.26,
"train_tokens_per_second": 4204.23
},
{
"epoch": 0.8519908023720199,
"grad_norm": 0.12209635825164061,
"learning_rate": 4.503843375410024e-05,
"loss": 0.0862,
"num_input_tokens_seen": 331011792,
"step": 880,
"train_runtime": 78589.3442,
"train_tokens_per_second": 4211.917
},
{
"epoch": 0.8568316592036791,
"grad_norm": 0.12814219418396117,
"learning_rate": 4.49539240785351e-05,
"loss": 0.0862,
"num_input_tokens_seen": 332759152,
"step": 885,
"train_runtime": 78840.0264,
"train_tokens_per_second": 4220.688
},
{
"epoch": 0.8616725160353382,
"grad_norm": 0.10182567961465826,
"learning_rate": 4.4868781452257604e-05,
"loss": 0.0854,
"num_input_tokens_seen": 334548544,
"step": 890,
"train_runtime": 79107.638,
"train_tokens_per_second": 4229.03
},
{
"epoch": 0.8665133728669975,
"grad_norm": 0.1059374934901486,
"learning_rate": 4.478300857604407e-05,
"loss": 0.081,
"num_input_tokens_seen": 336480736,
"step": 895,
"train_runtime": 79448.3723,
"train_tokens_per_second": 4235.212
},
{
"epoch": 0.8713542296986566,
"grad_norm": 0.11153870028757291,
"learning_rate": 4.469660817066277e-05,
"loss": 0.0799,
"num_input_tokens_seen": 338427280,
"step": 900,
"train_runtime": 79789.1685,
"train_tokens_per_second": 4241.519
},
{
"epoch": 0.8761950865303159,
"grad_norm": 0.13350987905341574,
"learning_rate": 4.460958297678756e-05,
"loss": 0.0828,
"num_input_tokens_seen": 340350320,
"step": 905,
"train_runtime": 80140.7635,
"train_tokens_per_second": 4246.906
},
{
"epoch": 0.8810359433619751,
"grad_norm": 0.10558711020322371,
"learning_rate": 4.452193575491096e-05,
"loss": 0.0815,
"num_input_tokens_seen": 342226288,
"step": 910,
"train_runtime": 80448.6307,
"train_tokens_per_second": 4253.973
},
{
"epoch": 0.8858768001936342,
"grad_norm": 0.1139445337081212,
"learning_rate": 4.443366928525663e-05,
"loss": 0.0813,
"num_input_tokens_seen": 344202544,
"step": 915,
"train_runtime": 80804.3894,
"train_tokens_per_second": 4259.701
},
{
"epoch": 0.8907176570252935,
"grad_norm": 0.12328404863523318,
"learning_rate": 4.4344786367691125e-05,
"loss": 0.0889,
"num_input_tokens_seen": 346094384,
"step": 920,
"train_runtime": 81120.3268,
"train_tokens_per_second": 4266.432
},
{
"epoch": 0.8955585138569527,
"grad_norm": 0.0973559450919178,
"learning_rate": 4.425528982163512e-05,
"loss": 0.0806,
"num_input_tokens_seen": 348009840,
"step": 925,
"train_runtime": 81458.3407,
"train_tokens_per_second": 4272.243
},
{
"epoch": 0.9003993706886119,
"grad_norm": 0.09888756400918433,
"learning_rate": 4.416518248597395e-05,
"loss": 0.0816,
"num_input_tokens_seen": 349923616,
"step": 930,
"train_runtime": 81791.2966,
"train_tokens_per_second": 4278.25
},
{
"epoch": 0.9052402275202711,
"grad_norm": 0.10273344580681865,
"learning_rate": 4.4074467218967594e-05,
"loss": 0.0808,
"num_input_tokens_seen": 351982800,
"step": 935,
"train_runtime": 82224.6329,
"train_tokens_per_second": 4280.746
},
{
"epoch": 0.9100810843519302,
"grad_norm": 0.1137919757065847,
"learning_rate": 4.398314689815995e-05,
"loss": 0.086,
"num_input_tokens_seen": 353803568,
"step": 940,
"train_runtime": 82523.3997,
"train_tokens_per_second": 4287.312
},
{
"epoch": 0.9149219411835895,
"grad_norm": 0.1129658121707594,
"learning_rate": 4.389122442028762e-05,
"loss": 0.0792,
"num_input_tokens_seen": 355770400,
"step": 945,
"train_runtime": 82863.0664,
"train_tokens_per_second": 4293.474
},
{
"epoch": 0.9197627980152487,
"grad_norm": 0.12295404647246533,
"learning_rate": 4.379870270118801e-05,
"loss": 0.0789,
"num_input_tokens_seen": 357722704,
"step": 950,
"train_runtime": 83211.7587,
"train_tokens_per_second": 4298.944
},
{
"epoch": 0.9246036548469079,
"grad_norm": 0.10702291365375034,
"learning_rate": 4.370558467570678e-05,
"loss": 0.0784,
"num_input_tokens_seen": 359625392,
"step": 955,
"train_runtime": 83503.0123,
"train_tokens_per_second": 4306.736
},
{
"epoch": 0.9294445116785671,
"grad_norm": 0.10671020858504487,
"learning_rate": 4.361187329760483e-05,
"loss": 0.0812,
"num_input_tokens_seen": 361478880,
"step": 960,
"train_runtime": 83796.3829,
"train_tokens_per_second": 4313.777
},
{
"epoch": 0.9342853685102263,
"grad_norm": 0.10376407616031118,
"learning_rate": 4.351757153946456e-05,
"loss": 0.0831,
"num_input_tokens_seen": 363401552,
"step": 965,
"train_runtime": 84125.1709,
"train_tokens_per_second": 4319.772
},
{
"epoch": 0.9391262253418855,
"grad_norm": 0.09840753061510307,
"learning_rate": 4.3422682392595594e-05,
"loss": 0.077,
"num_input_tokens_seen": 365256960,
"step": 970,
"train_runtime": 84396.4737,
"train_tokens_per_second": 4327.87
},
{
"epoch": 0.9439670821735447,
"grad_norm": 0.09244245286343136,
"learning_rate": 4.332720886693987e-05,
"loss": 0.0781,
"num_input_tokens_seen": 367167264,
"step": 975,
"train_runtime": 84737.6756,
"train_tokens_per_second": 4332.987
},
{
"epoch": 0.9488079390052039,
"grad_norm": 0.10909448781256045,
"learning_rate": 4.32311539909762e-05,
"loss": 0.084,
"num_input_tokens_seen": 368976208,
"step": 980,
"train_runtime": 85107.1009,
"train_tokens_per_second": 4335.434
},
{
"epoch": 0.9536487958368631,
"grad_norm": 0.10797216762579882,
"learning_rate": 4.313452081162416e-05,
"loss": 0.0829,
"num_input_tokens_seen": 370862208,
"step": 985,
"train_runtime": 85937.0527,
"train_tokens_per_second": 4315.51
},
{
"epoch": 0.9584896526685224,
"grad_norm": 0.09730637257808095,
"learning_rate": 4.303731239414749e-05,
"loss": 0.0859,
"num_input_tokens_seen": 372719648,
"step": 990,
"train_runtime": 86738.8585,
"train_tokens_per_second": 4297.032
},
{
"epoch": 0.9633305095001815,
"grad_norm": 0.10271769317718264,
"learning_rate": 4.2939531822056815e-05,
"loss": 0.0869,
"num_input_tokens_seen": 374624944,
"step": 995,
"train_runtime": 87558.6912,
"train_tokens_per_second": 4278.558
},
{
"epoch": 0.9681713663318408,
"grad_norm": 0.10573584364906888,
"learning_rate": 4.284118219701187e-05,
"loss": 0.0794,
"num_input_tokens_seen": 376525888,
"step": 1000,
"train_runtime": 88323.819,
"train_tokens_per_second": 4263.016
},
{
"epoch": 0.9730122231634999,
"grad_norm": 0.10710167799127884,
"learning_rate": 4.2742266638723096e-05,
"loss": 0.0804,
"num_input_tokens_seen": 378537840,
"step": 1005,
"train_runtime": 89245.7806,
"train_tokens_per_second": 4241.521
},
{
"epoch": 0.9778530799951591,
"grad_norm": 0.11197233290927573,
"learning_rate": 4.264278828485267e-05,
"loss": 0.0845,
"num_input_tokens_seen": 380397552,
"step": 1010,
"train_runtime": 90080.3151,
"train_tokens_per_second": 4222.871
},
{
"epoch": 0.9826939368268184,
"grad_norm": 0.11297920761471489,
"learning_rate": 4.254275029091501e-05,
"loss": 0.089,
"num_input_tokens_seen": 382191440,
"step": 1015,
"train_runtime": 90923.6166,
"train_tokens_per_second": 4203.434
},
{
"epoch": 0.9875347936584775,
"grad_norm": 0.11866337547317915,
"learning_rate": 4.2442155830176655e-05,
"loss": 0.0887,
"num_input_tokens_seen": 384056752,
"step": 1020,
"train_runtime": 91754.463,
"train_tokens_per_second": 4185.701
},
{
"epoch": 0.9923756504901368,
"grad_norm": 0.10420016956278523,
"learning_rate": 4.2341008093555604e-05,
"loss": 0.0833,
"num_input_tokens_seen": 385974032,
"step": 1025,
"train_runtime": 92603.2825,
"train_tokens_per_second": 4168.038
},
{
"epoch": 0.997216507321796,
"grad_norm": 0.11976582014616471,
"learning_rate": 4.22393102895201e-05,
"loss": 0.0807,
"num_input_tokens_seen": 387884720,
"step": 1030,
"train_runtime": 93427.3132,
"train_tokens_per_second": 4151.727
},
{
"epoch": 1.0019363427326637,
"grad_norm": 0.12104509350691006,
"learning_rate": 4.213706564398688e-05,
"loss": 0.0727,
"num_input_tokens_seen": 389685952,
"step": 1035,
"train_runtime": 94255.5336,
"train_tokens_per_second": 4134.356
},
{
"epoch": 1.0067771995643229,
"grad_norm": 0.10790601727390471,
"learning_rate": 4.203427740021884e-05,
"loss": 0.0713,
"num_input_tokens_seen": 391624192,
"step": 1040,
"train_runtime": 95169.1567,
"train_tokens_per_second": 4115.033
},
{
"epoch": 1.011618056395982,
"grad_norm": 0.09855811615223968,
"learning_rate": 4.1930948818722104e-05,
"loss": 0.0671,
"num_input_tokens_seen": 393660384,
"step": 1045,
"train_runtime": 96151.6752,
"train_tokens_per_second": 4094.16
},
{
"epoch": 1.0164589132276414,
"grad_norm": 0.10511029272951843,
"learning_rate": 4.182708317714267e-05,
"loss": 0.0686,
"num_input_tokens_seen": 395529008,
"step": 1050,
"train_runtime": 96960.8327,
"train_tokens_per_second": 4079.266
},
{
"epoch": 1.0212997700593005,
"grad_norm": 0.10834149804451805,
"learning_rate": 4.172268377016241e-05,
"loss": 0.0684,
"num_input_tokens_seen": 397347360,
"step": 1055,
"train_runtime": 97712.5511,
"train_tokens_per_second": 4066.493
},
{
"epoch": 1.0261406268909596,
"grad_norm": 0.09606298518108944,
"learning_rate": 4.161775390939454e-05,
"loss": 0.0664,
"num_input_tokens_seen": 399294608,
"step": 1060,
"train_runtime": 98616.012,
"train_tokens_per_second": 4048.984
},
{
"epoch": 1.030981483722619,
"grad_norm": 0.10742642762254552,
"learning_rate": 4.151229692327863e-05,
"loss": 0.0657,
"num_input_tokens_seen": 401232320,
"step": 1065,
"train_runtime": 99525.5766,
"train_tokens_per_second": 4031.449
},
{
"epoch": 1.0358223405542781,
"grad_norm": 0.099355578010755,
"learning_rate": 4.1406316156974965e-05,
"loss": 0.0668,
"num_input_tokens_seen": 403156576,
"step": 1070,
"train_runtime": 100492.0641,
"train_tokens_per_second": 4011.825
},
{
"epoch": 1.0406631973859373,
"grad_norm": 0.09750969581285164,
"learning_rate": 4.1299814972258466e-05,
"loss": 0.0685,
"num_input_tokens_seen": 404973808,
"step": 1075,
"train_runtime": 101227.6948,
"train_tokens_per_second": 4000.623
},
{
"epoch": 1.0455040542175964,
"grad_norm": 0.10948901174240346,
"learning_rate": 4.1192796747412046e-05,
"loss": 0.0739,
"num_input_tokens_seen": 406856480,
"step": 1080,
"train_runtime": 102069.5624,
"train_tokens_per_second": 3986.071
},
{
"epoch": 1.0503449110492558,
"grad_norm": 0.10315111502114589,
"learning_rate": 4.108526487711944e-05,
"loss": 0.0743,
"num_input_tokens_seen": 408659088,
"step": 1085,
"train_runtime": 102905.3681,
"train_tokens_per_second": 3971.213
},
{
"epoch": 1.055185767880915,
"grad_norm": 0.12136499885985239,
"learning_rate": 4.0977222772357545e-05,
"loss": 0.0674,
"num_input_tokens_seen": 410533120,
"step": 1090,
"train_runtime": 103694.499,
"train_tokens_per_second": 3959.064
},
{
"epoch": 1.060026624712574,
"grad_norm": 0.12011224246251123,
"learning_rate": 4.0868673860288196e-05,
"loss": 0.0723,
"num_input_tokens_seen": 412377808,
"step": 1095,
"train_runtime": 104483.7209,
"train_tokens_per_second": 3946.814
},
{
"epoch": 1.0648674815442334,
"grad_norm": 0.11631933894778271,
"learning_rate": 4.075962158414948e-05,
"loss": 0.0659,
"num_input_tokens_seen": 414390304,
"step": 1100,
"train_runtime": 105480.5472,
"train_tokens_per_second": 3928.595
},
{
"epoch": 1.0697083383758925,
"grad_norm": 0.10080944032903205,
"learning_rate": 4.0650069403146484e-05,
"loss": 0.0671,
"num_input_tokens_seen": 416413616,
"step": 1105,
"train_runtime": 106510.3272,
"train_tokens_per_second": 3909.608
},
{
"epoch": 1.0745491952075517,
"grad_norm": 0.09617342271272442,
"learning_rate": 4.0540020792341605e-05,
"loss": 0.0723,
"num_input_tokens_seen": 418267408,
"step": 1110,
"train_runtime": 107384.6318,
"train_tokens_per_second": 3895.04
},
{
"epoch": 1.079390052039211,
"grad_norm": 0.09814369288730117,
"learning_rate": 4.0429479242544285e-05,
"loss": 0.0676,
"num_input_tokens_seen": 420122000,
"step": 1115,
"train_runtime": 108179.5693,
"train_tokens_per_second": 3883.561
},
{
"epoch": 1.0842309088708701,
"grad_norm": 0.09566981500512636,
"learning_rate": 4.031844826020028e-05,
"loss": 0.0673,
"num_input_tokens_seen": 421979504,
"step": 1120,
"train_runtime": 108984.7112,
"train_tokens_per_second": 3871.915
},
{
"epoch": 1.0890717657025293,
"grad_norm": 0.11481624591628409,
"learning_rate": 4.020693136728047e-05,
"loss": 0.0716,
"num_input_tokens_seen": 423852128,
"step": 1125,
"train_runtime": 109849.6222,
"train_tokens_per_second": 3858.476
},
{
"epoch": 1.0939126225341886,
"grad_norm": 0.11408237955687323,
"learning_rate": 4.0094932101169116e-05,
"loss": 0.066,
"num_input_tokens_seen": 425865744,
"step": 1130,
"train_runtime": 110829.9796,
"train_tokens_per_second": 3842.514
},
{
"epoch": 1.0987534793658478,
"grad_norm": 0.09876683127617192,
"learning_rate": 3.9982454014551626e-05,
"loss": 0.0671,
"num_input_tokens_seen": 427825520,
"step": 1135,
"train_runtime": 111678.4669,
"train_tokens_per_second": 3830.868
},
{
"epoch": 1.103594336197507,
"grad_norm": 0.08998178840723907,
"learning_rate": 3.9869500675301925e-05,
"loss": 0.0659,
"num_input_tokens_seen": 429650464,
"step": 1140,
"train_runtime": 112452.9235,
"train_tokens_per_second": 3820.714
},
{
"epoch": 1.1084351930291663,
"grad_norm": 0.09112012767906727,
"learning_rate": 3.975607566636921e-05,
"loss": 0.0674,
"num_input_tokens_seen": 431560992,
"step": 1145,
"train_runtime": 113300.5965,
"train_tokens_per_second": 3808.991
},
{
"epoch": 1.1132760498608254,
"grad_norm": 0.09193978880262148,
"learning_rate": 3.964218258566436e-05,
"loss": 0.0679,
"num_input_tokens_seen": 433452176,
"step": 1150,
"train_runtime": 114121.3844,
"train_tokens_per_second": 3798.168
},
{
"epoch": 1.1181169066924845,
"grad_norm": 0.10253244394594298,
"learning_rate": 3.952782504594574e-05,
"loss": 0.0683,
"num_input_tokens_seen": 435371936,
"step": 1155,
"train_runtime": 114932.6591,
"train_tokens_per_second": 3788.061
},
{
"epoch": 1.1229577635241437,
"grad_norm": 0.11604949971136329,
"learning_rate": 3.9413006674704684e-05,
"loss": 0.0696,
"num_input_tokens_seen": 437191424,
"step": 1160,
"train_runtime": 115724.1081,
"train_tokens_per_second": 3777.877
},
{
"epoch": 1.127798620355803,
"grad_norm": 0.09600367492582994,
"learning_rate": 3.929773111405034e-05,
"loss": 0.0693,
"num_input_tokens_seen": 439045440,
"step": 1165,
"train_runtime": 116553.6206,
"train_tokens_per_second": 3766.897
},
{
"epoch": 1.1326394771874622,
"grad_norm": 0.10626811630147584,
"learning_rate": 3.9182002020594235e-05,
"loss": 0.0705,
"num_input_tokens_seen": 440950928,
"step": 1170,
"train_runtime": 117462.4396,
"train_tokens_per_second": 3753.974
},
{
"epoch": 1.1374803340191213,
"grad_norm": 0.09282379898506365,
"learning_rate": 3.906582306533418e-05,
"loss": 0.0638,
"num_input_tokens_seen": 442914960,
"step": 1175,
"train_runtime": 118455.494,
"train_tokens_per_second": 3739.083
},
{
"epoch": 1.1423211908507807,
"grad_norm": 0.09929776152503765,
"learning_rate": 3.8949197933537916e-05,
"loss": 0.0689,
"num_input_tokens_seen": 444832176,
"step": 1180,
"train_runtime": 119337.8677,
"train_tokens_per_second": 3727.502
},
{
"epoch": 1.1471620476824398,
"grad_norm": 0.1011962511269963,
"learning_rate": 3.883213032462617e-05,
"loss": 0.0682,
"num_input_tokens_seen": 446780448,
"step": 1185,
"train_runtime": 120254.4758,
"train_tokens_per_second": 3715.292
},
{
"epoch": 1.152002904514099,
"grad_norm": 0.09174519283988156,
"learning_rate": 3.871462395205531e-05,
"loss": 0.071,
"num_input_tokens_seen": 448550288,
"step": 1190,
"train_runtime": 121004.0348,
"train_tokens_per_second": 3706.904
},
{
"epoch": 1.1568437613457583,
"grad_norm": 0.10621267015218695,
"learning_rate": 3.8596682543199546e-05,
"loss": 0.068,
"num_input_tokens_seen": 450450928,
"step": 1195,
"train_runtime": 121845.7134,
"train_tokens_per_second": 3696.896
},
{
"epoch": 1.1616846181774174,
"grad_norm": 0.09664344676790702,
"learning_rate": 3.847830983923273e-05,
"loss": 0.0687,
"num_input_tokens_seen": 452419728,
"step": 1200,
"train_runtime": 122865.555,
"train_tokens_per_second": 3682.234
},
{
"epoch": 1.1665254750090766,
"grad_norm": 0.09223910440789163,
"learning_rate": 3.835950959500963e-05,
"loss": 0.0719,
"num_input_tokens_seen": 454259472,
"step": 1205,
"train_runtime": 123677.075,
"train_tokens_per_second": 3672.948
},
{
"epoch": 1.1713663318407357,
"grad_norm": 0.10736751066204356,
"learning_rate": 3.8240285578946904e-05,
"loss": 0.069,
"num_input_tokens_seen": 456165264,
"step": 1210,
"train_runtime": 124627.8308,
"train_tokens_per_second": 3660.22
},
{
"epoch": 1.176207188672395,
"grad_norm": 0.09115327170920039,
"learning_rate": 3.812064157290345e-05,
"loss": 0.0666,
"num_input_tokens_seen": 458044704,
"step": 1215,
"train_runtime": 125470.063,
"train_tokens_per_second": 3650.629
},
{
"epoch": 1.1810480455040542,
"grad_norm": 0.10235191270929159,
"learning_rate": 3.800058137206053e-05,
"loss": 0.0666,
"num_input_tokens_seen": 459916000,
"step": 1220,
"train_runtime": 126329.2794,
"train_tokens_per_second": 3640.613
},
{
"epoch": 1.1858889023357135,
"grad_norm": 0.11706669148345339,
"learning_rate": 3.788010878480139e-05,
"loss": 0.0721,
"num_input_tokens_seen": 461786304,
"step": 1225,
"train_runtime": 127251.258,
"train_tokens_per_second": 3628.933
},
{
"epoch": 1.1907297591673727,
"grad_norm": 0.09895227308048286,
"learning_rate": 3.775922763259038e-05,
"loss": 0.0661,
"num_input_tokens_seen": 463676192,
"step": 1230,
"train_runtime": 128161.1571,
"train_tokens_per_second": 3617.915
},
{
"epoch": 1.1955706159990318,
"grad_norm": 0.11484504551320268,
"learning_rate": 3.763794174985181e-05,
"loss": 0.0705,
"num_input_tokens_seen": 465602880,
"step": 1235,
"train_runtime": 129003.0382,
"train_tokens_per_second": 3609.24
},
{
"epoch": 1.200411472830691,
"grad_norm": 0.08994605167832181,
"learning_rate": 3.7516254983848277e-05,
"loss": 0.0698,
"num_input_tokens_seen": 467493856,
"step": 1240,
"train_runtime": 129744.0548,
"train_tokens_per_second": 3603.201
},
{
"epoch": 1.2052523296623503,
"grad_norm": 0.10300174734565927,
"learning_rate": 3.7394171194558646e-05,
"loss": 0.0693,
"num_input_tokens_seen": 469496768,
"step": 1245,
"train_runtime": 130459.977,
"train_tokens_per_second": 3598.78
},
{
"epoch": 1.2100931864940094,
"grad_norm": 0.10675389319087088,
"learning_rate": 3.727169425455562e-05,
"loss": 0.0693,
"num_input_tokens_seen": 471449056,
"step": 1250,
"train_runtime": 131010.3565,
"train_tokens_per_second": 3598.563
},
{
"epoch": 1.2149340433256686,
"grad_norm": 0.09921256550808967,
"learning_rate": 3.7148828048882857e-05,
"loss": 0.0697,
"num_input_tokens_seen": 473264848,
"step": 1255,
"train_runtime": 131606.2869,
"train_tokens_per_second": 3596.066
},
{
"epoch": 1.219774900157328,
"grad_norm": 0.10462373380428001,
"learning_rate": 3.702557647493177e-05,
"loss": 0.069,
"num_input_tokens_seen": 475144208,
"step": 1260,
"train_runtime": 132107.4645,
"train_tokens_per_second": 3596.649
},
{
"epoch": 1.224615756988987,
"grad_norm": 0.10415516508324393,
"learning_rate": 3.690194344231789e-05,
"loss": 0.0671,
"num_input_tokens_seen": 477118864,
"step": 1265,
"train_runtime": 132532.0609,
"train_tokens_per_second": 3600.026
},
{
"epoch": 1.2294566138206462,
"grad_norm": 0.10859351584240985,
"learning_rate": 3.677793287275687e-05,
"loss": 0.0697,
"num_input_tokens_seen": 478984544,
"step": 1270,
"train_runtime": 132834.8843,
"train_tokens_per_second": 3605.864
},
{
"epoch": 1.2342974706523056,
"grad_norm": 0.10820393055282612,
"learning_rate": 3.665354869994003e-05,
"loss": 0.073,
"num_input_tokens_seen": 480862384,
"step": 1275,
"train_runtime": 133150.7075,
"train_tokens_per_second": 3611.414
},
{
"epoch": 1.2391383274839647,
"grad_norm": 0.09874866443912496,
"learning_rate": 3.652879486940965e-05,
"loss": 0.071,
"num_input_tokens_seen": 482657712,
"step": 1280,
"train_runtime": 133461.2075,
"train_tokens_per_second": 3616.464
},
{
"epoch": 1.2439791843156238,
"grad_norm": 0.1063684888653233,
"learning_rate": 3.640367533843376e-05,
"loss": 0.0694,
"num_input_tokens_seen": 484527216,
"step": 1285,
"train_runtime": 133797.6333,
"train_tokens_per_second": 3621.344
},
{
"epoch": 1.2488200411472832,
"grad_norm": 0.11664187966473348,
"learning_rate": 3.6278194075880625e-05,
"loss": 0.0664,
"num_input_tokens_seen": 486424320,
"step": 1290,
"train_runtime": 134098.3231,
"train_tokens_per_second": 3627.371
},
{
"epoch": 1.2536608979789423,
"grad_norm": 0.08949284897230625,
"learning_rate": 3.615235506209288e-05,
"loss": 0.0686,
"num_input_tokens_seen": 488454832,
"step": 1295,
"train_runtime": 134472.7384,
"train_tokens_per_second": 3632.371
},
{
"epoch": 1.2585017548106014,
"grad_norm": 0.08869637799926437,
"learning_rate": 3.602616228876123e-05,
"loss": 0.0616,
"num_input_tokens_seen": 490406800,
"step": 1300,
"train_runtime": 134802.021,
"train_tokens_per_second": 3637.978
},
{
"epoch": 1.2633426116422606,
"grad_norm": 0.1058762928159983,
"learning_rate": 3.589961975879787e-05,
"loss": 0.0675,
"num_input_tokens_seen": 492319664,
"step": 1305,
"train_runtime": 135130.0473,
"train_tokens_per_second": 3643.303
},
{
"epoch": 1.26818346847392,
"grad_norm": 0.10187344379977778,
"learning_rate": 3.577273148620946e-05,
"loss": 0.0693,
"num_input_tokens_seen": 494216816,
"step": 1310,
"train_runtime": 135461.9383,
"train_tokens_per_second": 3648.381
},
{
"epoch": 1.273024325305579,
"grad_norm": 0.10989990396916802,
"learning_rate": 3.564550149596985e-05,
"loss": 0.0699,
"num_input_tokens_seen": 496046624,
"step": 1315,
"train_runtime": 135771.2575,
"train_tokens_per_second": 3653.547
},
{
"epoch": 1.2778651821372382,
"grad_norm": 0.10349545274745546,
"learning_rate": 3.5517933823892384e-05,
"loss": 0.0755,
"num_input_tokens_seen": 497876768,
"step": 1320,
"train_runtime": 136087.2124,
"train_tokens_per_second": 3658.513
},
{
"epoch": 1.2827060389688976,
"grad_norm": 0.09370049601860225,
"learning_rate": 3.539003251650188e-05,
"loss": 0.0786,
"num_input_tokens_seen": 499699664,
"step": 1325,
"train_runtime": 136409.5926,
"train_tokens_per_second": 3663.23
},
{
"epoch": 1.2875468958005567,
"grad_norm": 0.09622928750516618,
"learning_rate": 3.526180163090627e-05,
"loss": 0.0683,
"num_input_tokens_seen": 501536176,
"step": 1330,
"train_runtime": 136698.4884,
"train_tokens_per_second": 3668.923
},
{
"epoch": 1.2923877526322158,
"grad_norm": 0.09942000608374545,
"learning_rate": 3.5133245234667883e-05,
"loss": 0.0688,
"num_input_tokens_seen": 503368832,
"step": 1335,
"train_runtime": 137002.8399,
"train_tokens_per_second": 3674.149
},
{
"epoch": 1.297228609463875,
"grad_norm": 0.10312231028530282,
"learning_rate": 3.50043674056745e-05,
"loss": 0.0675,
"num_input_tokens_seen": 505269776,
"step": 1340,
"train_runtime": 137326.8406,
"train_tokens_per_second": 3679.323
},
{
"epoch": 1.3020694662955343,
"grad_norm": 0.10706010123105417,
"learning_rate": 3.48751722320099e-05,
"loss": 0.0646,
"num_input_tokens_seen": 507187600,
"step": 1345,
"train_runtime": 137656.0465,
"train_tokens_per_second": 3684.456
},
{
"epoch": 1.3069103231271935,
"grad_norm": 0.11190774683646115,
"learning_rate": 3.4745663811824234e-05,
"loss": 0.0701,
"num_input_tokens_seen": 509000368,
"step": 1350,
"train_runtime": 137932.7138,
"train_tokens_per_second": 3690.208
},
{
"epoch": 1.3117511799588528,
"grad_norm": 0.09559260030218834,
"learning_rate": 3.461584625320407e-05,
"loss": 0.0672,
"num_input_tokens_seen": 510812240,
"step": 1355,
"train_runtime": 138211.6321,
"train_tokens_per_second": 3695.87
},
{
"epoch": 1.316592036790512,
"grad_norm": 0.10592839240696805,
"learning_rate": 3.448572367404197e-05,
"loss": 0.0753,
"num_input_tokens_seen": 512638608,
"step": 1360,
"train_runtime": 138535.165,
"train_tokens_per_second": 3700.422
},
{
"epoch": 1.321432893622171,
"grad_norm": 0.10547925559174358,
"learning_rate": 3.435530020190598e-05,
"loss": 0.0666,
"num_input_tokens_seen": 514458464,
"step": 1365,
"train_runtime": 138805.6716,
"train_tokens_per_second": 3706.322
},
{
"epoch": 1.3262737504538302,
"grad_norm": 0.10826688240565788,
"learning_rate": 3.422457997390865e-05,
"loss": 0.0675,
"num_input_tokens_seen": 516280880,
"step": 1370,
"train_runtime": 139079.93,
"train_tokens_per_second": 3712.116
},
{
"epoch": 1.3311146072854896,
"grad_norm": 0.10770450004766066,
"learning_rate": 3.4093567136575794e-05,
"loss": 0.0726,
"num_input_tokens_seen": 518154288,
"step": 1375,
"train_runtime": 139387.3129,
"train_tokens_per_second": 3717.371
},
{
"epoch": 1.3359554641171487,
"grad_norm": 0.1142302841713288,
"learning_rate": 3.396226584571499e-05,
"loss": 0.0747,
"num_input_tokens_seen": 519986144,
"step": 1380,
"train_runtime": 139690.3762,
"train_tokens_per_second": 3722.419
},
{
"epoch": 1.340796320948808,
"grad_norm": 0.1002246744241518,
"learning_rate": 3.383068026628371e-05,
"loss": 0.07,
"num_input_tokens_seen": 521862240,
"step": 1385,
"train_runtime": 139987.7519,
"train_tokens_per_second": 3727.914
},
{
"epoch": 1.3456371777804672,
"grad_norm": 0.09964010260687982,
"learning_rate": 3.3698814572257284e-05,
"loss": 0.0665,
"num_input_tokens_seen": 523858656,
"step": 1390,
"train_runtime": 140338.2763,
"train_tokens_per_second": 3732.828
},
{
"epoch": 1.3504780346121263,
"grad_norm": 0.09090502017139336,
"learning_rate": 3.356667294649639e-05,
"loss": 0.0675,
"num_input_tokens_seen": 525774560,
"step": 1395,
"train_runtime": 140671.1613,
"train_tokens_per_second": 3737.614
},
{
"epoch": 1.3553188914437855,
"grad_norm": 0.10302633457655123,
"learning_rate": 3.343425958061447e-05,
"loss": 0.0662,
"num_input_tokens_seen": 527681344,
"step": 1400,
"train_runtime": 140994.6669,
"train_tokens_per_second": 3742.562
},
{
"epoch": 1.3601597482754448,
"grad_norm": 0.09470780975547861,
"learning_rate": 3.330157867484472e-05,
"loss": 0.0717,
"num_input_tokens_seen": 529406736,
"step": 1405,
"train_runtime": 141236.0937,
"train_tokens_per_second": 3748.381
},
{
"epoch": 1.365000605107104,
"grad_norm": 0.08585499478944604,
"learning_rate": 3.3168634437906865e-05,
"loss": 0.0662,
"num_input_tokens_seen": 531305872,
"step": 1410,
"train_runtime": 141547.5703,
"train_tokens_per_second": 3753.55
},
{
"epoch": 1.369841461938763,
"grad_norm": 0.10732846763853142,
"learning_rate": 3.303543108687364e-05,
"loss": 0.0715,
"num_input_tokens_seen": 533037392,
"step": 1415,
"train_runtime": 141806.9679,
"train_tokens_per_second": 3758.894
},
{
"epoch": 1.3746823187704225,
"grad_norm": 0.09819540974635672,
"learning_rate": 3.290197284703707e-05,
"loss": 0.0687,
"num_input_tokens_seen": 535085280,
"step": 1420,
"train_runtime": 142243.4525,
"train_tokens_per_second": 3761.757
},
{
"epoch": 1.3795231756020816,
"grad_norm": 0.09528337700443731,
"learning_rate": 3.276826395177438e-05,
"loss": 0.0728,
"num_input_tokens_seen": 536966576,
"step": 1425,
"train_runtime": 142532.1226,
"train_tokens_per_second": 3767.337
},
{
"epoch": 1.3843640324337407,
"grad_norm": 0.09402247417532383,
"learning_rate": 3.263430864241376e-05,
"loss": 0.068,
"num_input_tokens_seen": 538829216,
"step": 1430,
"train_runtime": 142855.1544,
"train_tokens_per_second": 3771.857
},
{
"epoch": 1.3892048892653999,
"grad_norm": 0.09890632658294073,
"learning_rate": 3.250011116809978e-05,
"loss": 0.0678,
"num_input_tokens_seen": 540632144,
"step": 1435,
"train_runtime": 143161.7888,
"train_tokens_per_second": 3776.372
},
{
"epoch": 1.3940457460970592,
"grad_norm": 0.08589478003550063,
"learning_rate": 3.236567578565867e-05,
"loss": 0.0693,
"num_input_tokens_seen": 542432288,
"step": 1440,
"train_runtime": 143420.8223,
"train_tokens_per_second": 3782.103
},
{
"epoch": 1.3988866029287184,
"grad_norm": 0.09609704465838043,
"learning_rate": 3.223100675946321e-05,
"loss": 0.0655,
"num_input_tokens_seen": 544403440,
"step": 1445,
"train_runtime": 143809.3419,
"train_tokens_per_second": 3785.592
},
{
"epoch": 1.4037274597603777,
"grad_norm": 0.08644550430034957,
"learning_rate": 3.209610836129755e-05,
"loss": 0.0657,
"num_input_tokens_seen": 546192304,
"step": 1450,
"train_runtime": 144057.4772,
"train_tokens_per_second": 3791.489
},
{
"epoch": 1.4085683165920369,
"grad_norm": 0.10119133528568032,
"learning_rate": 3.1960984870221596e-05,
"loss": 0.065,
"num_input_tokens_seen": 548030912,
"step": 1455,
"train_runtime": 144334.6546,
"train_tokens_per_second": 3796.946
},
{
"epoch": 1.413409173423696,
"grad_norm": 0.09755652421098708,
"learning_rate": 3.1825640572435394e-05,
"loss": 0.067,
"num_input_tokens_seen": 549968368,
"step": 1460,
"train_runtime": 144647.7235,
"train_tokens_per_second": 3802.123
},
{
"epoch": 1.4182500302553551,
"grad_norm": 0.09202887102207528,
"learning_rate": 3.169007976114311e-05,
"loss": 0.0633,
"num_input_tokens_seen": 551997376,
"step": 1465,
"train_runtime": 145066.4375,
"train_tokens_per_second": 3805.135
},
{
"epoch": 1.4230908870870145,
"grad_norm": 0.09893125478101567,
"learning_rate": 3.155430673641681e-05,
"loss": 0.069,
"num_input_tokens_seen": 553815552,
"step": 1470,
"train_runtime": 145350.8539,
"train_tokens_per_second": 3810.198
},
{
"epoch": 1.4279317439186736,
"grad_norm": 0.11231020588498852,
"learning_rate": 3.1418325805060126e-05,
"loss": 0.0713,
"num_input_tokens_seen": 555719392,
"step": 1475,
"train_runtime": 145699.1273,
"train_tokens_per_second": 3814.157
},
{
"epoch": 1.4327726007503327,
"grad_norm": 0.10568451565662428,
"learning_rate": 3.1282141280471645e-05,
"loss": 0.069,
"num_input_tokens_seen": 557504016,
"step": 1480,
"train_runtime": 145974.2182,
"train_tokens_per_second": 3819.195
},
{
"epoch": 1.437613457581992,
"grad_norm": 0.09645253967773308,
"learning_rate": 3.114575748250801e-05,
"loss": 0.0678,
"num_input_tokens_seen": 559452496,
"step": 1485,
"train_runtime": 146328.064,
"train_tokens_per_second": 3823.275
},
{
"epoch": 1.4424543144136512,
"grad_norm": 0.09105344674720453,
"learning_rate": 3.100917873734696e-05,
"loss": 0.0648,
"num_input_tokens_seen": 561337136,
"step": 1490,
"train_runtime": 146656.3116,
"train_tokens_per_second": 3827.569
},
{
"epoch": 1.4472951712453104,
"grad_norm": 0.09122782894664445,
"learning_rate": 3.087240937735008e-05,
"loss": 0.0682,
"num_input_tokens_seen": 563237008,
"step": 1495,
"train_runtime": 146953.3037,
"train_tokens_per_second": 3832.762
},
{
"epoch": 1.4521360280769695,
"grad_norm": 0.10094808912146574,
"learning_rate": 3.073545374092535e-05,
"loss": 0.07,
"num_input_tokens_seen": 565056704,
"step": 1500,
"train_runtime": 147252.9238,
"train_tokens_per_second": 3837.321
},
{
"epoch": 1.4569768849086289,
"grad_norm": 0.09898552497334528,
"learning_rate": 3.05983161723896e-05,
"loss": 0.0653,
"num_input_tokens_seen": 566984576,
"step": 1505,
"train_runtime": 147575.0281,
"train_tokens_per_second": 3842.009
},
{
"epoch": 1.461817741740288,
"grad_norm": 0.0893045303914291,
"learning_rate": 3.046100102183061e-05,
"loss": 0.0664,
"num_input_tokens_seen": 568931872,
"step": 1510,
"train_runtime": 147890.4027,
"train_tokens_per_second": 3846.983
},
{
"epoch": 1.4666585985719474,
"grad_norm": 0.07794027254932458,
"learning_rate": 3.0323512644969194e-05,
"loss": 0.0644,
"num_input_tokens_seen": 570766304,
"step": 1515,
"train_runtime": 148153.9685,
"train_tokens_per_second": 3852.521
},
{
"epoch": 1.4714994554036065,
"grad_norm": 0.09387987802719616,
"learning_rate": 3.0185855403021013e-05,
"loss": 0.0697,
"num_input_tokens_seen": 572692384,
"step": 1520,
"train_runtime": 148490.2566,
"train_tokens_per_second": 3856.767
},
{
"epoch": 1.4763403122352656,
"grad_norm": 0.10537208908005448,
"learning_rate": 3.0048033662558222e-05,
"loss": 0.0731,
"num_input_tokens_seen": 574478256,
"step": 1525,
"train_runtime": 148778.7811,
"train_tokens_per_second": 3861.292
},
{
"epoch": 1.4811811690669248,
"grad_norm": 0.09520737006467728,
"learning_rate": 2.9910051795370974e-05,
"loss": 0.068,
"num_input_tokens_seen": 576337936,
"step": 1530,
"train_runtime": 149086.1454,
"train_tokens_per_second": 3865.805
},
{
"epoch": 1.4860220258985841,
"grad_norm": 0.09478631735802089,
"learning_rate": 2.977191417832874e-05,
"loss": 0.0663,
"num_input_tokens_seen": 578104656,
"step": 1535,
"train_runtime": 149330.6478,
"train_tokens_per_second": 3871.306
},
{
"epoch": 1.4908628827302433,
"grad_norm": 0.10196404573493925,
"learning_rate": 2.9633625193241475e-05,
"loss": 0.0674,
"num_input_tokens_seen": 579956016,
"step": 1540,
"train_runtime": 149621.4808,
"train_tokens_per_second": 3876.155
},
{
"epoch": 1.4957037395619024,
"grad_norm": 0.10883093122161902,
"learning_rate": 2.9495189226720615e-05,
"loss": 0.0694,
"num_input_tokens_seen": 581873728,
"step": 1545,
"train_runtime": 149975.2694,
"train_tokens_per_second": 3879.798
},
{
"epoch": 1.5005445963935617,
"grad_norm": 0.09938067661414723,
"learning_rate": 2.935661067003994e-05,
"loss": 0.0689,
"num_input_tokens_seen": 583722192,
"step": 1550,
"train_runtime": 150285.3755,
"train_tokens_per_second": 3884.092
},
{
"epoch": 1.5053854532252209,
"grad_norm": 0.10557220766786837,
"learning_rate": 2.9217893918996285e-05,
"loss": 0.0711,
"num_input_tokens_seen": 585531280,
"step": 1555,
"train_runtime": 150574.9739,
"train_tokens_per_second": 3888.636
},
{
"epoch": 1.51022631005688,
"grad_norm": 0.09792652092850067,
"learning_rate": 2.9079043373770088e-05,
"loss": 0.0657,
"num_input_tokens_seen": 587465120,
"step": 1560,
"train_runtime": 150916.7039,
"train_tokens_per_second": 3892.645
},
{
"epoch": 1.5150671668885392,
"grad_norm": 0.10432359138697463,
"learning_rate": 2.8940063438785808e-05,
"loss": 0.065,
"num_input_tokens_seen": 589312272,
"step": 1565,
"train_runtime": 151233.2823,
"train_tokens_per_second": 3896.71
},
{
"epoch": 1.5199080237201985,
"grad_norm": 0.10464650554060877,
"learning_rate": 2.8800958522572246e-05,
"loss": 0.0669,
"num_input_tokens_seen": 591224496,
"step": 1570,
"train_runtime": 151542.6129,
"train_tokens_per_second": 3901.375
},
{
"epoch": 1.5247488805518576,
"grad_norm": 0.08982894126025302,
"learning_rate": 2.866173303762268e-05,
"loss": 0.0703,
"num_input_tokens_seen": 593094000,
"step": 1575,
"train_runtime": 151870.8088,
"train_tokens_per_second": 3905.253
},
{
"epoch": 1.529589737383517,
"grad_norm": 0.09170883604725243,
"learning_rate": 2.8522391400254887e-05,
"loss": 0.0678,
"num_input_tokens_seen": 595038832,
"step": 1580,
"train_runtime": 152202.7319,
"train_tokens_per_second": 3909.515
},
{
"epoch": 1.5344305942151761,
"grad_norm": 0.10514358953867606,
"learning_rate": 2.8382938030471112e-05,
"loss": 0.0697,
"num_input_tokens_seen": 596824960,
"step": 1585,
"train_runtime": 152452.6889,
"train_tokens_per_second": 3914.821
},
{
"epoch": 1.5392714510468353,
"grad_norm": 0.10505102469174953,
"learning_rate": 2.8243377351817755e-05,
"loss": 0.0689,
"num_input_tokens_seen": 598705200,
"step": 1590,
"train_runtime": 152787.4796,
"train_tokens_per_second": 3918.549
},
{
"epoch": 1.5441123078784944,
"grad_norm": 0.09749265072972622,
"learning_rate": 2.8103713791245178e-05,
"loss": 0.0612,
"num_input_tokens_seen": 600600592,
"step": 1595,
"train_runtime": 153095.0077,
"train_tokens_per_second": 3923.058
},
{
"epoch": 1.5489531647101535,
"grad_norm": 0.09875360040532327,
"learning_rate": 2.7963951778967197e-05,
"loss": 0.0681,
"num_input_tokens_seen": 602389504,
"step": 1600,
"train_runtime": 153336.2956,
"train_tokens_per_second": 3928.551
},
{
"epoch": 1.553794021541813,
"grad_norm": 0.0943003245484352,
"learning_rate": 2.7824095748320568e-05,
"loss": 0.0697,
"num_input_tokens_seen": 604104288,
"step": 1605,
"train_runtime": 153582.7315,
"train_tokens_per_second": 3933.413
},
{
"epoch": 1.5586348783734723,
"grad_norm": 0.08913383289373572,
"learning_rate": 2.7684150135624376e-05,
"loss": 0.0701,
"num_input_tokens_seen": 605951504,
"step": 1610,
"train_runtime": 153880.8942,
"train_tokens_per_second": 3937.796
},
{
"epoch": 1.5634757352051314,
"grad_norm": 0.09806492578910929,
"learning_rate": 2.7544119380039314e-05,
"loss": 0.0661,
"num_input_tokens_seen": 607990208,
"step": 1615,
"train_runtime": 154299.517,
"train_tokens_per_second": 3940.325
},
{
"epoch": 1.5683165920367905,
"grad_norm": 0.09468491378509171,
"learning_rate": 2.740400792342685e-05,
"loss": 0.0699,
"num_input_tokens_seen": 609909232,
"step": 1620,
"train_runtime": 154619.4097,
"train_tokens_per_second": 3944.584
},
{
"epoch": 1.5731574488684497,
"grad_norm": 0.11209084242527069,
"learning_rate": 2.726382021020833e-05,
"loss": 0.0718,
"num_input_tokens_seen": 611736000,
"step": 1625,
"train_runtime": 154922.3873,
"train_tokens_per_second": 3948.661
},
{
"epoch": 1.5779983057001088,
"grad_norm": 0.08630941935012056,
"learning_rate": 2.7123560687224013e-05,
"loss": 0.0665,
"num_input_tokens_seen": 613559536,
"step": 1630,
"train_runtime": 155208.4781,
"train_tokens_per_second": 3953.132
},
{
"epoch": 1.5828391625317682,
"grad_norm": 0.0933989919648124,
"learning_rate": 2.6983233803592022e-05,
"loss": 0.0664,
"num_input_tokens_seen": 615391184,
"step": 1635,
"train_runtime": 155485.9117,
"train_tokens_per_second": 3957.858
},
{
"epoch": 1.5876800193634273,
"grad_norm": 0.08973133951027497,
"learning_rate": 2.6842844010567177e-05,
"loss": 0.0734,
"num_input_tokens_seen": 617305296,
"step": 1640,
"train_runtime": 155858.1926,
"train_tokens_per_second": 3960.686
},
{
"epoch": 1.5925208761950866,
"grad_norm": 0.0933153070243224,
"learning_rate": 2.6702395761399844e-05,
"loss": 0.0674,
"num_input_tokens_seen": 619107456,
"step": 1645,
"train_runtime": 156128.9329,
"train_tokens_per_second": 3965.36
},
{
"epoch": 1.5973617330267458,
"grad_norm": 0.10662169623290256,
"learning_rate": 2.656189351119463e-05,
"loss": 0.0705,
"num_input_tokens_seen": 620950304,
"step": 1650,
"train_runtime": 156466.5875,
"train_tokens_per_second": 3968.581
},
{
"epoch": 1.602202589858405,
"grad_norm": 0.10372114619240043,
"learning_rate": 2.6421341716769112e-05,
"loss": 0.0692,
"num_input_tokens_seen": 622796352,
"step": 1655,
"train_runtime": 156767.9586,
"train_tokens_per_second": 3972.727
},
{
"epoch": 1.607043446690064,
"grad_norm": 0.09132496975982439,
"learning_rate": 2.6280744836512423e-05,
"loss": 0.0666,
"num_input_tokens_seen": 624698480,
"step": 1660,
"train_runtime": 157088.7012,
"train_tokens_per_second": 3976.724
},
{
"epoch": 1.6118843035217232,
"grad_norm": 0.09728380085896536,
"learning_rate": 2.6140107330243858e-05,
"loss": 0.0696,
"num_input_tokens_seen": 626526800,
"step": 1665,
"train_runtime": 157381.2207,
"train_tokens_per_second": 3980.95
},
{
"epoch": 1.6167251603533825,
"grad_norm": 0.10083788994513473,
"learning_rate": 2.599943365907138e-05,
"loss": 0.0708,
"num_input_tokens_seen": 628340752,
"step": 1670,
"train_runtime": 157653.9021,
"train_tokens_per_second": 3985.571
},
{
"epoch": 1.621566017185042,
"grad_norm": 0.09206012606935819,
"learning_rate": 2.5858728285250156e-05,
"loss": 0.0628,
"num_input_tokens_seen": 630294736,
"step": 1675,
"train_runtime": 157993.6687,
"train_tokens_per_second": 3989.367
},
{
"epoch": 1.626406874016701,
"grad_norm": 0.11407194887653826,
"learning_rate": 2.5717995672040934e-05,
"loss": 0.0668,
"num_input_tokens_seen": 632181360,
"step": 1680,
"train_runtime": 158306.6385,
"train_tokens_per_second": 3993.398
},
{
"epoch": 1.6312477308483602,
"grad_norm": 0.10532893116463045,
"learning_rate": 2.5577240283568547e-05,
"loss": 0.0672,
"num_input_tokens_seen": 634139120,
"step": 1685,
"train_runtime": 158659.1124,
"train_tokens_per_second": 3996.865
},
{
"epoch": 1.6360885876800193,
"grad_norm": 0.09253705708958208,
"learning_rate": 2.5436466584680257e-05,
"loss": 0.065,
"num_input_tokens_seen": 636061136,
"step": 1690,
"train_runtime": 158996.3128,
"train_tokens_per_second": 4000.477
},
{
"epoch": 1.6409294445116784,
"grad_norm": 0.10270850201788718,
"learning_rate": 2.529567904080416e-05,
"loss": 0.0679,
"num_input_tokens_seen": 637901488,
"step": 1695,
"train_runtime": 159300.0136,
"train_tokens_per_second": 4004.403
},
{
"epoch": 1.6457703013433378,
"grad_norm": 0.1020950381213344,
"learning_rate": 2.5154882117807503e-05,
"loss": 0.0674,
"num_input_tokens_seen": 639783152,
"step": 1700,
"train_runtime": 159641.5796,
"train_tokens_per_second": 4007.622
},
{
"epoch": 1.650611158174997,
"grad_norm": 0.10434335557369012,
"learning_rate": 2.5014080281855062e-05,
"loss": 0.0718,
"num_input_tokens_seen": 641601136,
"step": 1705,
"train_runtime": 159948.6783,
"train_tokens_per_second": 4011.294
},
{
"epoch": 1.6554520150066563,
"grad_norm": 0.09269120806836094,
"learning_rate": 2.4873277999267443e-05,
"loss": 0.0682,
"num_input_tokens_seen": 643464448,
"step": 1710,
"train_runtime": 160228.7859,
"train_tokens_per_second": 4015.91
},
{
"epoch": 1.6602928718383154,
"grad_norm": 0.09545551967743877,
"learning_rate": 2.473247973637942e-05,
"loss": 0.0668,
"num_input_tokens_seen": 645403984,
"step": 1715,
"train_runtime": 160573.5465,
"train_tokens_per_second": 4019.367
},
{
"epoch": 1.6651337286699746,
"grad_norm": 0.09245805131119629,
"learning_rate": 2.459168995939827e-05,
"loss": 0.063,
"num_input_tokens_seen": 647359520,
"step": 1720,
"train_runtime": 160916.6879,
"train_tokens_per_second": 4022.948
},
{
"epoch": 1.6699745855016337,
"grad_norm": 0.1130522027501556,
"learning_rate": 2.4450913134262077e-05,
"loss": 0.0698,
"num_input_tokens_seen": 649221056,
"step": 1725,
"train_runtime": 161212.5554,
"train_tokens_per_second": 4027.112
},
{
"epoch": 1.674815442333293,
"grad_norm": 0.1001979401813176,
"learning_rate": 2.4310153726498107e-05,
"loss": 0.0713,
"num_input_tokens_seen": 651074128,
"step": 1730,
"train_runtime": 161567.7558,
"train_tokens_per_second": 4029.728
},
{
"epoch": 1.6796562991649522,
"grad_norm": 0.09664199798231568,
"learning_rate": 2.4169416201081102e-05,
"loss": 0.0697,
"num_input_tokens_seen": 652978208,
"step": 1735,
"train_runtime": 161942.97,
"train_tokens_per_second": 4032.149
},
{
"epoch": 1.6844971559966115,
"grad_norm": 0.10551245269293205,
"learning_rate": 2.4028705022291728e-05,
"loss": 0.0661,
"num_input_tokens_seen": 654910336,
"step": 1740,
"train_runtime": 162300.8039,
"train_tokens_per_second": 4035.164
},
{
"epoch": 1.6893380128282707,
"grad_norm": 0.0926086339529829,
"learning_rate": 2.3888024653574865e-05,
"loss": 0.0654,
"num_input_tokens_seen": 656834320,
"step": 1745,
"train_runtime": 162632.724,
"train_tokens_per_second": 4038.759
},
{
"epoch": 1.6941788696599298,
"grad_norm": 0.0997990172837255,
"learning_rate": 2.374737955739814e-05,
"loss": 0.0675,
"num_input_tokens_seen": 658744624,
"step": 1750,
"train_runtime": 162977.7639,
"train_tokens_per_second": 4041.929
},
{
"epoch": 1.699019726491589,
"grad_norm": 0.09727233240178633,
"learning_rate": 2.3606774195110256e-05,
"loss": 0.0688,
"num_input_tokens_seen": 660657152,
"step": 1755,
"train_runtime": 163311.2124,
"train_tokens_per_second": 4045.388
},
{
"epoch": 1.703860583323248,
"grad_norm": 0.09186313556651109,
"learning_rate": 2.346621302679957e-05,
"loss": 0.0626,
"num_input_tokens_seen": 662618160,
"step": 1760,
"train_runtime": 163650.2492,
"train_tokens_per_second": 4048.99
},
{
"epoch": 1.7087014401549074,
"grad_norm": 0.09815622832004638,
"learning_rate": 2.3325700511152572e-05,
"loss": 0.0688,
"num_input_tokens_seen": 664466176,
"step": 1765,
"train_runtime": 163980.8046,
"train_tokens_per_second": 4052.097
},
{
"epoch": 1.7135422969865666,
"grad_norm": 0.09851131213295354,
"learning_rate": 2.3185241105312435e-05,
"loss": 0.0644,
"num_input_tokens_seen": 666303008,
"step": 1770,
"train_runtime": 164258.4151,
"train_tokens_per_second": 4056.431
},
{
"epoch": 1.718383153818226,
"grad_norm": 0.09650945128005722,
"learning_rate": 2.3044839264737695e-05,
"loss": 0.0654,
"num_input_tokens_seen": 668249808,
"step": 1775,
"train_runtime": 164606.7007,
"train_tokens_per_second": 4059.676
},
{
"epoch": 1.723224010649885,
"grad_norm": 0.07976228580722033,
"learning_rate": 2.2904499443060852e-05,
"loss": 0.0618,
"num_input_tokens_seen": 670142528,
"step": 1780,
"train_runtime": 164916.446,
"train_tokens_per_second": 4063.528
},
{
"epoch": 1.7280648674815442,
"grad_norm": 0.09316901052660637,
"learning_rate": 2.2764226091947153e-05,
"loss": 0.0705,
"num_input_tokens_seen": 671892176,
"step": 1785,
"train_runtime": 165204.9592,
"train_tokens_per_second": 4067.022
},
{
"epoch": 1.7329057243132033,
"grad_norm": 0.10554377156841822,
"learning_rate": 2.2624023660953322e-05,
"loss": 0.0694,
"num_input_tokens_seen": 673790848,
"step": 1790,
"train_runtime": 165555.898,
"train_tokens_per_second": 4069.869
},
{
"epoch": 1.7377465811448627,
"grad_norm": 0.1000010423520895,
"learning_rate": 2.2483896597386506e-05,
"loss": 0.0672,
"num_input_tokens_seen": 675763984,
"step": 1795,
"train_runtime": 165926.0539,
"train_tokens_per_second": 4072.682
},
{
"epoch": 1.7425874379765218,
"grad_norm": 0.09335668641339016,
"learning_rate": 2.2343849346163092e-05,
"loss": 0.0664,
"num_input_tokens_seen": 677686256,
"step": 1800,
"train_runtime": 166242.0359,
"train_tokens_per_second": 4076.504
},
{
"epoch": 1.7474282948081812,
"grad_norm": 0.09333896876761687,
"learning_rate": 2.2203886349667826e-05,
"loss": 0.067,
"num_input_tokens_seen": 679548000,
"step": 1805,
"train_runtime": 166529.1496,
"train_tokens_per_second": 4080.655
},
{
"epoch": 1.7522691516398403,
"grad_norm": 0.11018536147162415,
"learning_rate": 2.2064012047612796e-05,
"loss": 0.0674,
"num_input_tokens_seen": 681323824,
"step": 1810,
"train_runtime": 166764.5916,
"train_tokens_per_second": 4085.542
},
{
"epoch": 1.7571100084714995,
"grad_norm": 0.09634536795886925,
"learning_rate": 2.1924230876896684e-05,
"loss": 0.0639,
"num_input_tokens_seen": 683289152,
"step": 1815,
"train_runtime": 167115.3972,
"train_tokens_per_second": 4088.726
},
{
"epoch": 1.7619508653031586,
"grad_norm": 0.079098965621037,
"learning_rate": 2.1784547271463966e-05,
"loss": 0.063,
"num_input_tokens_seen": 685248960,
"step": 1820,
"train_runtime": 167461.2987,
"train_tokens_per_second": 4091.984
},
{
"epoch": 1.7667917221348177,
"grad_norm": 0.09262520602448539,
"learning_rate": 2.164496566216428e-05,
"loss": 0.0658,
"num_input_tokens_seen": 687075264,
"step": 1825,
"train_runtime": 167728.4786,
"train_tokens_per_second": 4096.354
},
{
"epoch": 1.771632578966477,
"grad_norm": 0.08425246771738228,
"learning_rate": 2.150549047661193e-05,
"loss": 0.0706,
"num_input_tokens_seen": 688930992,
"step": 1830,
"train_runtime": 168041.5415,
"train_tokens_per_second": 4099.766
},
{
"epoch": 1.7764734357981364,
"grad_norm": 0.08261309626975141,
"learning_rate": 2.136612613904533e-05,
"loss": 0.0633,
"num_input_tokens_seen": 690736016,
"step": 1835,
"train_runtime": 168299.5399,
"train_tokens_per_second": 4104.206
},
{
"epoch": 1.7813142926297956,
"grad_norm": 0.08439774632959215,
"learning_rate": 2.1226877070186782e-05,
"loss": 0.068,
"num_input_tokens_seen": 692670496,
"step": 1840,
"train_runtime": 168637.0439,
"train_tokens_per_second": 4107.463
},
{
"epoch": 1.7861551494614547,
"grad_norm": 0.09065583044936652,
"learning_rate": 2.108774768710215e-05,
"loss": 0.0643,
"num_input_tokens_seen": 694613552,
"step": 1845,
"train_runtime": 168980.2344,
"train_tokens_per_second": 4110.62
},
{
"epoch": 1.7909960062931138,
"grad_norm": 0.09792211257284066,
"learning_rate": 2.0948742403060838e-05,
"loss": 0.0667,
"num_input_tokens_seen": 696469216,
"step": 1850,
"train_runtime": 169280.592,
"train_tokens_per_second": 4114.289
},
{
"epoch": 1.795836863124773,
"grad_norm": 0.09897529531743958,
"learning_rate": 2.0809865627395705e-05,
"loss": 0.0681,
"num_input_tokens_seen": 698418624,
"step": 1855,
"train_runtime": 169653.3494,
"train_tokens_per_second": 4116.739
},
{
"epoch": 1.8006777199564323,
"grad_norm": 0.09500857629505935,
"learning_rate": 2.0671121765363288e-05,
"loss": 0.0647,
"num_input_tokens_seen": 700384480,
"step": 1860,
"train_runtime": 170043.8506,
"train_tokens_per_second": 4118.846
},
{
"epoch": 1.8055185767880915,
"grad_norm": 0.10247249755147927,
"learning_rate": 2.0532515218003985e-05,
"loss": 0.0699,
"num_input_tokens_seen": 702304704,
"step": 1865,
"train_runtime": 170404.8081,
"train_tokens_per_second": 4121.39
},
{
"epoch": 1.8103594336197508,
"grad_norm": 0.09643799643651368,
"learning_rate": 2.039405038200252e-05,
"loss": 0.0691,
"num_input_tokens_seen": 704195280,
"step": 1870,
"train_runtime": 170715.849,
"train_tokens_per_second": 4124.955
},
{
"epoch": 1.81520029045141,
"grad_norm": 0.09857814992539433,
"learning_rate": 2.0255731649548427e-05,
"loss": 0.0656,
"num_input_tokens_seen": 706059712,
"step": 1875,
"train_runtime": 171033.6245,
"train_tokens_per_second": 4128.192
},
{
"epoch": 1.820041147283069,
"grad_norm": 0.09418823224621559,
"learning_rate": 2.011756340819673e-05,
"loss": 0.0649,
"num_input_tokens_seen": 708049376,
"step": 1880,
"train_runtime": 171420.7357,
"train_tokens_per_second": 4130.477
},
{
"epoch": 1.8248820041147282,
"grad_norm": 0.10339589373318703,
"learning_rate": 1.9979550040728838e-05,
"loss": 0.0677,
"num_input_tokens_seen": 709957648,
"step": 1885,
"train_runtime": 171751.8988,
"train_tokens_per_second": 4133.623
},
{
"epoch": 1.8297228609463874,
"grad_norm": 0.09532254902656445,
"learning_rate": 1.9841695925013406e-05,
"loss": 0.0666,
"num_input_tokens_seen": 711775040,
"step": 1890,
"train_runtime": 172045.6813,
"train_tokens_per_second": 4137.128
},
{
"epoch": 1.8345637177780467,
"grad_norm": 0.0880459196194006,
"learning_rate": 1.9704005433867555e-05,
"loss": 0.0652,
"num_input_tokens_seen": 713667296,
"step": 1895,
"train_runtime": 172364.1363,
"train_tokens_per_second": 4140.463
},
{
"epoch": 1.839404574609706,
"grad_norm": 0.08969514913899801,
"learning_rate": 1.9566482934918135e-05,
"loss": 0.0653,
"num_input_tokens_seen": 715572784,
"step": 1900,
"train_runtime": 172680.2824,
"train_tokens_per_second": 4143.917
},
{
"epoch": 1.8442454314413652,
"grad_norm": 0.0913565853305896,
"learning_rate": 1.9429132790463172e-05,
"loss": 0.0623,
"num_input_tokens_seen": 717403184,
"step": 1905,
"train_runtime": 172952.4861,
"train_tokens_per_second": 4147.978
},
{
"epoch": 1.8490862882730243,
"grad_norm": 0.07640087073283276,
"learning_rate": 1.9291959357333495e-05,
"loss": 0.0652,
"num_input_tokens_seen": 719188672,
"step": 1910,
"train_runtime": 173218.7348,
"train_tokens_per_second": 4151.91
},
{
"epoch": 1.8539271451046835,
"grad_norm": 0.08816630513451665,
"learning_rate": 1.9154966986754568e-05,
"loss": 0.0656,
"num_input_tokens_seen": 721142368,
"step": 1915,
"train_runtime": 173569.3058,
"train_tokens_per_second": 4154.781
},
{
"epoch": 1.8587680019363426,
"grad_norm": 0.09071657241594651,
"learning_rate": 1.9018160024208406e-05,
"loss": 0.0673,
"num_input_tokens_seen": 722942560,
"step": 1920,
"train_runtime": 173839.4391,
"train_tokens_per_second": 4158.68
},
{
"epoch": 1.863608858768002,
"grad_norm": 0.10455029456382117,
"learning_rate": 1.8881542809295804e-05,
"loss": 0.0659,
"num_input_tokens_seen": 724852000,
"step": 1925,
"train_runtime": 174168.4044,
"train_tokens_per_second": 4161.788
},
{
"epoch": 1.868449715599661,
"grad_norm": 0.08954830475655517,
"learning_rate": 1.874511967559861e-05,
"loss": 0.0635,
"num_input_tokens_seen": 726715728,
"step": 1930,
"train_runtime": 174457.6742,
"train_tokens_per_second": 4165.57
},
{
"epoch": 1.8732905724313205,
"grad_norm": 0.09323369420411323,
"learning_rate": 1.860889495054231e-05,
"loss": 0.0675,
"num_input_tokens_seen": 728548704,
"step": 1935,
"train_runtime": 174741.255,
"train_tokens_per_second": 4169.3
},
{
"epoch": 1.8781314292629796,
"grad_norm": 0.09793727462020756,
"learning_rate": 1.8472872955258764e-05,
"loss": 0.065,
"num_input_tokens_seen": 730456800,
"step": 1940,
"train_runtime": 175068.9574,
"train_tokens_per_second": 4172.395
},
{
"epoch": 1.8829722860946387,
"grad_norm": 0.09583230673463962,
"learning_rate": 1.8337058004449087e-05,
"loss": 0.066,
"num_input_tokens_seen": 732296016,
"step": 1945,
"train_runtime": 175374.5228,
"train_tokens_per_second": 4175.612
},
{
"epoch": 1.8878131429262979,
"grad_norm": 0.09835179058080876,
"learning_rate": 1.8201454406246853e-05,
"loss": 0.0651,
"num_input_tokens_seen": 734259984,
"step": 1950,
"train_runtime": 175740.5312,
"train_tokens_per_second": 4178.091
},
{
"epoch": 1.892653999757957,
"grad_norm": 0.09108637153342643,
"learning_rate": 1.8066066462081365e-05,
"loss": 0.0626,
"num_input_tokens_seen": 736134608,
"step": 1955,
"train_runtime": 176053.0546,
"train_tokens_per_second": 4181.323
},
{
"epoch": 1.8974948565896164,
"grad_norm": 0.0919055897190125,
"learning_rate": 1.7930898466541278e-05,
"loss": 0.0643,
"num_input_tokens_seen": 737950480,
"step": 1960,
"train_runtime": 176349.4191,
"train_tokens_per_second": 4184.593
},
{
"epoch": 1.9023357134212757,
"grad_norm": 0.090159941398429,
"learning_rate": 1.779595470723831e-05,
"loss": 0.0615,
"num_input_tokens_seen": 739837152,
"step": 1965,
"train_runtime": 176648.1234,
"train_tokens_per_second": 4188.197
},
{
"epoch": 1.9071765702529349,
"grad_norm": 0.09346294098357426,
"learning_rate": 1.7661239464671307e-05,
"loss": 0.0665,
"num_input_tokens_seen": 741721280,
"step": 1970,
"train_runtime": 176965.8973,
"train_tokens_per_second": 4191.323
},
{
"epoch": 1.912017427084594,
"grad_norm": 0.10241633630793719,
"learning_rate": 1.7526757012090384e-05,
"loss": 0.0651,
"num_input_tokens_seen": 743570432,
"step": 1975,
"train_runtime": 177269.1725,
"train_tokens_per_second": 4194.584
},
{
"epoch": 1.9168582839162531,
"grad_norm": 0.09173533340918434,
"learning_rate": 1.7392511615361454e-05,
"loss": 0.0651,
"num_input_tokens_seen": 745332208,
"step": 1980,
"train_runtime": 177532.7109,
"train_tokens_per_second": 4198.281
},
{
"epoch": 1.9216991407479123,
"grad_norm": 0.1008623059871893,
"learning_rate": 1.7258507532830843e-05,
"loss": 0.0648,
"num_input_tokens_seen": 747317616,
"step": 1985,
"train_runtime": 177898.9302,
"train_tokens_per_second": 4200.799
},
{
"epoch": 1.9265399975795716,
"grad_norm": 0.089928131700152,
"learning_rate": 1.7124749015190245e-05,
"loss": 0.0646,
"num_input_tokens_seen": 749180048,
"step": 1990,
"train_runtime": 178198.867,
"train_tokens_per_second": 4204.18
},
{
"epoch": 1.9313808544112308,
"grad_norm": 0.08994654227932986,
"learning_rate": 1.699124030534191e-05,
"loss": 0.0629,
"num_input_tokens_seen": 751039536,
"step": 1995,
"train_runtime": 178494.6526,
"train_tokens_per_second": 4207.63
},
{
"epoch": 1.93622171124289,
"grad_norm": 0.08906162646309361,
"learning_rate": 1.6857985638263994e-05,
"loss": 0.0669,
"num_input_tokens_seen": 752911840,
"step": 2000,
"train_runtime": 178774.0756,
"train_tokens_per_second": 4211.527
},
{
"epoch": 1.9410625680745492,
"grad_norm": 0.0871511140825133,
"learning_rate": 1.6724989240876302e-05,
"loss": 0.0633,
"num_input_tokens_seen": 754784096,
"step": 2005,
"train_runtime": 179064.7161,
"train_tokens_per_second": 4215.147
},
{
"epoch": 1.9459034249062084,
"grad_norm": 0.09886328312958423,
"learning_rate": 1.6592255331906127e-05,
"loss": 0.0631,
"num_input_tokens_seen": 756649472,
"step": 2010,
"train_runtime": 179383.4531,
"train_tokens_per_second": 4218.056
},
{
"epoch": 1.9507442817378675,
"grad_norm": 0.09426446935032833,
"learning_rate": 1.64597881217545e-05,
"loss": 0.0611,
"num_input_tokens_seen": 758536336,
"step": 2015,
"train_runtime": 179682.0449,
"train_tokens_per_second": 4221.548
},
{
"epoch": 1.9555851385695266,
"grad_norm": 0.09397155441601662,
"learning_rate": 1.632759181236258e-05,
"loss": 0.0653,
"num_input_tokens_seen": 760377888,
"step": 2020,
"train_runtime": 179989.7183,
"train_tokens_per_second": 4224.563
},
{
"epoch": 1.960425995401186,
"grad_norm": 0.10368768607789577,
"learning_rate": 1.61956705970784e-05,
"loss": 0.0674,
"num_input_tokens_seen": 762278528,
"step": 2025,
"train_runtime": 180338.7056,
"train_tokens_per_second": 4226.927
},
{
"epoch": 1.9652668522328454,
"grad_norm": 0.0931039697726731,
"learning_rate": 1.606402866052382e-05,
"loss": 0.0666,
"num_input_tokens_seen": 764061840,
"step": 2030,
"train_runtime": 180601.1349,
"train_tokens_per_second": 4230.659
},
{
"epoch": 1.9701077090645045,
"grad_norm": 0.09153549712253775,
"learning_rate": 1.593267017846185e-05,
"loss": 0.0652,
"num_input_tokens_seen": 765994112,
"step": 2035,
"train_runtime": 180938.3007,
"train_tokens_per_second": 4233.455
},
{
"epoch": 1.9749485658961636,
"grad_norm": 0.0910008764482439,
"learning_rate": 1.5801599317664106e-05,
"loss": 0.0655,
"num_input_tokens_seen": 767872592,
"step": 2040,
"train_runtime": 181245.9015,
"train_tokens_per_second": 4236.634
},
{
"epoch": 1.9797894227278228,
"grad_norm": 0.09525002435832676,
"learning_rate": 1.567082023577869e-05,
"loss": 0.0643,
"num_input_tokens_seen": 769751856,
"step": 2045,
"train_runtime": 181570.7397,
"train_tokens_per_second": 4239.405
},
{
"epoch": 1.984630279559482,
"grad_norm": 0.08882860457825476,
"learning_rate": 1.554033708119832e-05,
"loss": 0.0614,
"num_input_tokens_seen": 771662720,
"step": 2050,
"train_runtime": 181876.0901,
"train_tokens_per_second": 4242.794
},
{
"epoch": 1.9894711363911413,
"grad_norm": 0.09856363331292063,
"learning_rate": 1.5410153992928685e-05,
"loss": 0.0637,
"num_input_tokens_seen": 773536048,
"step": 2055,
"train_runtime": 182203.6654,
"train_tokens_per_second": 4245.447
},
{
"epoch": 1.9943119932228004,
"grad_norm": 0.09210117173688946,
"learning_rate": 1.528027510045723e-05,
"loss": 0.0647,
"num_input_tokens_seen": 775448176,
"step": 2060,
"train_runtime": 182525.7774,
"train_tokens_per_second": 4248.431
},
{
"epoch": 1.9991528500544598,
"grad_norm": 0.09915830664554921,
"learning_rate": 1.5150704523622066e-05,
"loss": 0.0633,
"num_input_tokens_seen": 777334320,
"step": 2065,
"train_runtime": 182819.8586,
"train_tokens_per_second": 4251.914
},
{
"epoch": 2.0038726854653275,
"grad_norm": 0.08510468955619002,
"learning_rate": 1.5021446372481408e-05,
"loss": 0.0456,
"num_input_tokens_seen": 779269280,
"step": 2070,
"train_runtime": 183203.0122,
"train_tokens_per_second": 4253.583
},
{
"epoch": 2.0087135422969866,
"grad_norm": 0.11048348864635747,
"learning_rate": 1.4892504747183078e-05,
"loss": 0.0469,
"num_input_tokens_seen": 781110080,
"step": 2075,
"train_runtime": 183493.0172,
"train_tokens_per_second": 4256.893
},
{
"epoch": 2.0135543991286458,
"grad_norm": 0.08898938959054886,
"learning_rate": 1.4763883737834538e-05,
"loss": 0.0424,
"num_input_tokens_seen": 783014480,
"step": 2080,
"train_runtime": 183815.8349,
"train_tokens_per_second": 4259.777
},
{
"epoch": 2.018395255960305,
"grad_norm": 0.0919658947768258,
"learning_rate": 1.4635587424373104e-05,
"loss": 0.0432,
"num_input_tokens_seen": 784918160,
"step": 2085,
"train_runtime": 184106.3305,
"train_tokens_per_second": 4263.396
},
{
"epoch": 2.023236112791964,
"grad_norm": 0.09232021098074011,
"learning_rate": 1.4507619876436548e-05,
"loss": 0.0417,
"num_input_tokens_seen": 786925872,
"step": 2090,
"train_runtime": 184467.9736,
"train_tokens_per_second": 4265.921
},
{
"epoch": 2.0280769696236236,
"grad_norm": 0.09272252103462124,
"learning_rate": 1.4379985153233983e-05,
"loss": 0.0427,
"num_input_tokens_seen": 788883568,
"step": 2095,
"train_runtime": 184855.3063,
"train_tokens_per_second": 4267.573
},
{
"epoch": 2.0329178264552827,
"grad_norm": 0.08716923303447359,
"learning_rate": 1.4252687303417125e-05,
"loss": 0.0447,
"num_input_tokens_seen": 790834704,
"step": 2100,
"train_runtime": 185222.062,
"train_tokens_per_second": 4269.657
},
{
"epoch": 2.037758683286942,
"grad_norm": 0.09674916113938142,
"learning_rate": 1.4125730364951867e-05,
"loss": 0.044,
"num_input_tokens_seen": 792644976,
"step": 2105,
"train_runtime": 185490.7671,
"train_tokens_per_second": 4273.231
},
{
"epoch": 2.042599540118601,
"grad_norm": 0.1018255197888474,
"learning_rate": 1.3999118364990172e-05,
"loss": 0.0432,
"num_input_tokens_seen": 794521776,
"step": 2110,
"train_runtime": 185814.5824,
"train_tokens_per_second": 4275.885
},
{
"epoch": 2.04744039695026,
"grad_norm": 0.10641267828109918,
"learning_rate": 1.387285531974235e-05,
"loss": 0.0465,
"num_input_tokens_seen": 796324144,
"step": 2115,
"train_runtime": 186093.9032,
"train_tokens_per_second": 4279.152
},
{
"epoch": 2.0522812537819193,
"grad_norm": 0.08890413404478513,
"learning_rate": 1.3746945234349651e-05,
"loss": 0.042,
"num_input_tokens_seen": 798176336,
"step": 2120,
"train_runtime": 186381.3552,
"train_tokens_per_second": 4282.49
},
{
"epoch": 2.0571221106135784,
"grad_norm": 0.1007973258825116,
"learning_rate": 1.3621392102757236e-05,
"loss": 0.0454,
"num_input_tokens_seen": 799981648,
"step": 2125,
"train_runtime": 186688.6329,
"train_tokens_per_second": 4285.112
},
{
"epoch": 2.061962967445238,
"grad_norm": 0.11828945277047198,
"learning_rate": 1.3496199907587426e-05,
"loss": 0.0447,
"num_input_tokens_seen": 801919920,
"step": 2130,
"train_runtime": 187027.4012,
"train_tokens_per_second": 4287.714
},
{
"epoch": 2.066803824276897,
"grad_norm": 0.09697782214954971,
"learning_rate": 1.337137262001349e-05,
"loss": 0.0422,
"num_input_tokens_seen": 803830048,
"step": 2135,
"train_runtime": 187371.5567,
"train_tokens_per_second": 4290.032
},
{
"epoch": 2.0716446811085563,
"grad_norm": 0.08971480039994223,
"learning_rate": 1.324691419963352e-05,
"loss": 0.0414,
"num_input_tokens_seen": 805707856,
"step": 2140,
"train_runtime": 187658.5794,
"train_tokens_per_second": 4293.477
},
{
"epoch": 2.0764855379402154,
"grad_norm": 0.09701860651683053,
"learning_rate": 1.3122828594345007e-05,
"loss": 0.0445,
"num_input_tokens_seen": 807627408,
"step": 2145,
"train_runtime": 187990.6441,
"train_tokens_per_second": 4296.104
},
{
"epoch": 2.0813263947718745,
"grad_norm": 0.10086672081682901,
"learning_rate": 1.2999119740219434e-05,
"loss": 0.0462,
"num_input_tokens_seen": 809504384,
"step": 2150,
"train_runtime": 188275.3974,
"train_tokens_per_second": 4299.576
},
{
"epoch": 2.0861672516035337,
"grad_norm": 0.10444978322400048,
"learning_rate": 1.2875791561377557e-05,
"loss": 0.0458,
"num_input_tokens_seen": 811379696,
"step": 2155,
"train_runtime": 188599.6081,
"train_tokens_per_second": 4302.128
},
{
"epoch": 2.091008108435193,
"grad_norm": 0.10161748986408596,
"learning_rate": 1.2752847969864857e-05,
"loss": 0.0425,
"num_input_tokens_seen": 813165728,
"step": 2160,
"train_runtime": 188873.2381,
"train_tokens_per_second": 4305.352
},
{
"epoch": 2.0958489652668524,
"grad_norm": 0.09934783137965864,
"learning_rate": 1.2630292865527483e-05,
"loss": 0.0425,
"num_input_tokens_seen": 815080720,
"step": 2165,
"train_runtime": 189177.9808,
"train_tokens_per_second": 4308.539
},
{
"epoch": 2.1006898220985115,
"grad_norm": 0.09905949770742634,
"learning_rate": 1.2508130135888518e-05,
"loss": 0.0425,
"num_input_tokens_seen": 817073936,
"step": 2170,
"train_runtime": 189630.3128,
"train_tokens_per_second": 4308.773
},
{
"epoch": 2.1055306789301707,
"grad_norm": 0.09067487917664219,
"learning_rate": 1.2386363656024691e-05,
"loss": 0.0415,
"num_input_tokens_seen": 818989136,
"step": 2175,
"train_runtime": 189949.6186,
"train_tokens_per_second": 4311.612
},
{
"epoch": 2.11037153576183,
"grad_norm": 0.10406493878646128,
"learning_rate": 1.226499728844345e-05,
"loss": 0.0425,
"num_input_tokens_seen": 820972848,
"step": 2180,
"train_runtime": 190334.3589,
"train_tokens_per_second": 4313.319
},
{
"epoch": 2.115212392593489,
"grad_norm": 0.08540798243716224,
"learning_rate": 1.2144034882960392e-05,
"loss": 0.0398,
"num_input_tokens_seen": 822903904,
"step": 2185,
"train_runtime": 190672.7343,
"train_tokens_per_second": 4315.792
},
{
"epoch": 2.120053249425148,
"grad_norm": 0.11741293089028593,
"learning_rate": 1.2023480276577267e-05,
"loss": 0.0473,
"num_input_tokens_seen": 824814944,
"step": 2190,
"train_runtime": 191027.2891,
"train_tokens_per_second": 4317.786
},
{
"epoch": 2.1248941062568076,
"grad_norm": 0.09836368383563045,
"learning_rate": 1.1903337293360101e-05,
"loss": 0.042,
"num_input_tokens_seen": 826652352,
"step": 2195,
"train_runtime": 191295.2085,
"train_tokens_per_second": 4321.344
},
{
"epoch": 2.1297349630884668,
"grad_norm": 0.1091691815964545,
"learning_rate": 1.1783609744318057e-05,
"loss": 0.0419,
"num_input_tokens_seen": 828587888,
"step": 2200,
"train_runtime": 191647.5762,
"train_tokens_per_second": 4323.498
},
{
"epoch": 2.134575819920126,
"grad_norm": 0.0943679758427896,
"learning_rate": 1.1664301427282418e-05,
"loss": 0.0449,
"num_input_tokens_seen": 830391184,
"step": 2205,
"train_runtime": 191919.3897,
"train_tokens_per_second": 4326.771
},
{
"epoch": 2.139416676751785,
"grad_norm": 0.10612562433762897,
"learning_rate": 1.154541612678619e-05,
"loss": 0.0476,
"num_input_tokens_seen": 832282752,
"step": 2210,
"train_runtime": 192247.4177,
"train_tokens_per_second": 4329.227
},
{
"epoch": 2.144257533583444,
"grad_norm": 0.09834530066415154,
"learning_rate": 1.1426957613944039e-05,
"loss": 0.0457,
"num_input_tokens_seen": 834136992,
"step": 2215,
"train_runtime": 192568.1333,
"train_tokens_per_second": 4331.646
},
{
"epoch": 2.1490983904151033,
"grad_norm": 0.11010939625022646,
"learning_rate": 1.1308929646332658e-05,
"loss": 0.0417,
"num_input_tokens_seen": 836052320,
"step": 2220,
"train_runtime": 192890.3051,
"train_tokens_per_second": 4334.341
},
{
"epoch": 2.153939247246763,
"grad_norm": 0.0989548927712895,
"learning_rate": 1.1191335967871582e-05,
"loss": 0.0439,
"num_input_tokens_seen": 837885568,
"step": 2225,
"train_runtime": 193169.477,
"train_tokens_per_second": 4337.567
},
{
"epoch": 2.158780104078422,
"grad_norm": 0.10316511914427559,
"learning_rate": 1.1074180308704429e-05,
"loss": 0.0448,
"num_input_tokens_seen": 839786048,
"step": 2230,
"train_runtime": 193482.1859,
"train_tokens_per_second": 4340.379
},
{
"epoch": 2.163620960910081,
"grad_norm": 0.11447985050959848,
"learning_rate": 1.0957466385080578e-05,
"loss": 0.0462,
"num_input_tokens_seen": 841636432,
"step": 2235,
"train_runtime": 193776.8505,
"train_tokens_per_second": 4343.328
},
{
"epoch": 2.1684618177417403,
"grad_norm": 0.09641942816445281,
"learning_rate": 1.0841197899237282e-05,
"loss": 0.0435,
"num_input_tokens_seen": 843588608,
"step": 2240,
"train_runtime": 194131.362,
"train_tokens_per_second": 4345.452
},
{
"epoch": 2.1733026745733994,
"grad_norm": 0.08526832186648046,
"learning_rate": 1.0725378539282246e-05,
"loss": 0.0435,
"num_input_tokens_seen": 845476688,
"step": 2245,
"train_runtime": 194465.4517,
"train_tokens_per_second": 4347.696
},
{
"epoch": 2.1781435314050586,
"grad_norm": 0.09389266894701039,
"learning_rate": 1.061001197907659e-05,
"loss": 0.0426,
"num_input_tokens_seen": 847341792,
"step": 2250,
"train_runtime": 194755.117,
"train_tokens_per_second": 4350.806
},
{
"epoch": 2.1829843882367177,
"grad_norm": 0.1091155029328961,
"learning_rate": 1.0495101878118408e-05,
"loss": 0.0438,
"num_input_tokens_seen": 849140496,
"step": 2255,
"train_runtime": 195018.4821,
"train_tokens_per_second": 4354.154
},
{
"epoch": 2.1878252450683773,
"grad_norm": 0.10639940814249475,
"learning_rate": 1.0380651881426567e-05,
"loss": 0.0421,
"num_input_tokens_seen": 850942480,
"step": 2260,
"train_runtime": 195260.0607,
"train_tokens_per_second": 4357.996
},
{
"epoch": 2.1926661019000364,
"grad_norm": 0.10716301854067384,
"learning_rate": 1.0266665619425181e-05,
"loss": 0.0441,
"num_input_tokens_seen": 852750800,
"step": 2265,
"train_runtime": 195553.3537,
"train_tokens_per_second": 4360.707
},
{
"epoch": 2.1975069587316955,
"grad_norm": 0.10343872393389002,
"learning_rate": 1.0153146707828404e-05,
"loss": 0.0427,
"num_input_tokens_seen": 854703456,
"step": 2270,
"train_runtime": 195905.746,
"train_tokens_per_second": 4362.83
},
{
"epoch": 2.2023478155633547,
"grad_norm": 0.09659246514397754,
"learning_rate": 1.004009874752575e-05,
"loss": 0.0438,
"num_input_tokens_seen": 856635792,
"step": 2275,
"train_runtime": 196280.3262,
"train_tokens_per_second": 4364.349
},
{
"epoch": 2.207188672395014,
"grad_norm": 0.10837299321798222,
"learning_rate": 9.927525324467866e-06,
"loss": 0.0435,
"num_input_tokens_seen": 858544176,
"step": 2280,
"train_runtime": 196631.0128,
"train_tokens_per_second": 4366.27
},
{
"epoch": 2.212029529226673,
"grad_norm": 0.0935449758044754,
"learning_rate": 9.81543000955279e-06,
"loss": 0.0412,
"num_input_tokens_seen": 860409424,
"step": 2285,
"train_runtime": 196926.666,
"train_tokens_per_second": 4369.187
},
{
"epoch": 2.2168703860583325,
"grad_norm": 0.08817931437627177,
"learning_rate": 9.703816358512674e-06,
"loss": 0.0415,
"num_input_tokens_seen": 862320048,
"step": 2290,
"train_runtime": 197238.6013,
"train_tokens_per_second": 4371.964
},
{
"epoch": 2.2217112428899917,
"grad_norm": 0.0980673558112737,
"learning_rate": 9.592687911801001e-06,
"loss": 0.0442,
"num_input_tokens_seen": 864243216,
"step": 2295,
"train_runtime": 197594.875,
"train_tokens_per_second": 4373.814
},
{
"epoch": 2.226552099721651,
"grad_norm": 0.10853728697423164,
"learning_rate": 9.482048194480275e-06,
"loss": 0.0425,
"num_input_tokens_seen": 866168848,
"step": 2300,
"train_runtime": 197933.6478,
"train_tokens_per_second": 4376.057
},
{
"epoch": 2.23139295655331,
"grad_norm": 0.11490821008247167,
"learning_rate": 9.371900716110183e-06,
"loss": 0.0431,
"num_input_tokens_seen": 868041296,
"step": 2305,
"train_runtime": 198239.5933,
"train_tokens_per_second": 4378.748
},
{
"epoch": 2.236233813384969,
"grad_norm": 0.08947053295422262,
"learning_rate": 9.262248970636334e-06,
"loss": 0.046,
"num_input_tokens_seen": 869878384,
"step": 2310,
"train_runtime": 198532.6139,
"train_tokens_per_second": 4381.539
},
{
"epoch": 2.241074670216628,
"grad_norm": 0.10485732709538129,
"learning_rate": 9.153096436279327e-06,
"loss": 0.042,
"num_input_tokens_seen": 871688688,
"step": 2315,
"train_runtime": 198817.926,
"train_tokens_per_second": 4384.357
},
{
"epoch": 2.2459155270482873,
"grad_norm": 0.11044109934649306,
"learning_rate": 9.044446575424548e-06,
"loss": 0.0431,
"num_input_tokens_seen": 873728640,
"step": 2320,
"train_runtime": 199228.9285,
"train_tokens_per_second": 4385.551
},
{
"epoch": 2.250756383879947,
"grad_norm": 0.09950408440238417,
"learning_rate": 8.936302834512206e-06,
"loss": 0.0439,
"num_input_tokens_seen": 875601696,
"step": 2325,
"train_runtime": 199550.3994,
"train_tokens_per_second": 4387.872
},
{
"epoch": 2.255597240711606,
"grad_norm": 0.09167335947382782,
"learning_rate": 8.828668643928112e-06,
"loss": 0.0419,
"num_input_tokens_seen": 877559760,
"step": 2330,
"train_runtime": 199908.7951,
"train_tokens_per_second": 4389.801
},
{
"epoch": 2.260438097543265,
"grad_norm": 0.10598202575372626,
"learning_rate": 8.721547417894816e-06,
"loss": 0.0432,
"num_input_tokens_seen": 879392576,
"step": 2335,
"train_runtime": 200216.2117,
"train_tokens_per_second": 4392.215
},
{
"epoch": 2.2652789543749243,
"grad_norm": 0.10543835687376972,
"learning_rate": 8.614942554363326e-06,
"loss": 0.0453,
"num_input_tokens_seen": 881180800,
"step": 2340,
"train_runtime": 200501.3334,
"train_tokens_per_second": 4394.887
},
{
"epoch": 2.2701198112065835,
"grad_norm": 0.10496418814685284,
"learning_rate": 8.508857434905304e-06,
"loss": 0.0423,
"num_input_tokens_seen": 883048928,
"step": 2345,
"train_runtime": 200822.2237,
"train_tokens_per_second": 4397.167
},
{
"epoch": 2.2749606680382426,
"grad_norm": 0.09577382338976835,
"learning_rate": 8.40329542460582e-06,
"loss": 0.0424,
"num_input_tokens_seen": 885000208,
"step": 2350,
"train_runtime": 201157.0354,
"train_tokens_per_second": 4399.549
},
{
"epoch": 2.279801524869902,
"grad_norm": 0.10452053840305,
"learning_rate": 8.298259871956596e-06,
"loss": 0.0432,
"num_input_tokens_seen": 886939040,
"step": 2355,
"train_runtime": 201495.1483,
"train_tokens_per_second": 4401.789
},
{
"epoch": 2.2846423817015613,
"grad_norm": 0.08543950174023542,
"learning_rate": 8.193754108749793e-06,
"loss": 0.0408,
"num_input_tokens_seen": 888957616,
"step": 2360,
"train_runtime": 201854.2624,
"train_tokens_per_second": 4403.958
},
{
"epoch": 2.2894832385332204,
"grad_norm": 0.10273943575259807,
"learning_rate": 8.08978144997234e-06,
"loss": 0.0427,
"num_input_tokens_seen": 890755008,
"step": 2365,
"train_runtime": 202132.5142,
"train_tokens_per_second": 4406.787
},
{
"epoch": 2.2943240953648796,
"grad_norm": 0.10320774278040343,
"learning_rate": 7.986345193700726e-06,
"loss": 0.0414,
"num_input_tokens_seen": 892675136,
"step": 2370,
"train_runtime": 202472.4309,
"train_tokens_per_second": 4408.873
},
{
"epoch": 2.2991649521965387,
"grad_norm": 0.10693768647967408,
"learning_rate": 7.883448620996489e-06,
"loss": 0.0427,
"num_input_tokens_seen": 894455920,
"step": 2375,
"train_runtime": 202743.8126,
"train_tokens_per_second": 4411.754
},
{
"epoch": 2.304005809028198,
"grad_norm": 0.10326302338915108,
"learning_rate": 7.781094995802007e-06,
"loss": 0.0485,
"num_input_tokens_seen": 896339664,
"step": 2380,
"train_runtime": 203093.6325,
"train_tokens_per_second": 4413.431
},
{
"epoch": 2.308846665859857,
"grad_norm": 0.1056092406773901,
"learning_rate": 7.679287564837074e-06,
"loss": 0.0417,
"num_input_tokens_seen": 898166624,
"step": 2385,
"train_runtime": 203393.0523,
"train_tokens_per_second": 4415.916
},
{
"epoch": 2.3136875226915166,
"grad_norm": 0.10998323658142867,
"learning_rate": 7.57802955749585e-06,
"loss": 0.0408,
"num_input_tokens_seen": 900181312,
"step": 2390,
"train_runtime": 203758.3229,
"train_tokens_per_second": 4417.887
},
{
"epoch": 2.3185283795231757,
"grad_norm": 0.1005967471563839,
"learning_rate": 7.4773241857444415e-06,
"loss": 0.0423,
"num_input_tokens_seen": 902067280,
"step": 2395,
"train_runtime": 204090.2987,
"train_tokens_per_second": 4419.942
},
{
"epoch": 2.323369236354835,
"grad_norm": 0.09602599680101273,
"learning_rate": 7.3771746440190195e-06,
"loss": 0.0417,
"num_input_tokens_seen": 903961952,
"step": 2400,
"train_runtime": 204390.7435,
"train_tokens_per_second": 4422.715
},
{
"epoch": 2.328210093186494,
"grad_norm": 0.10546331216246121,
"learning_rate": 7.277584109124483e-06,
"loss": 0.0439,
"num_input_tokens_seen": 905764752,
"step": 2405,
"train_runtime": 204689.5804,
"train_tokens_per_second": 4425.065
},
{
"epoch": 2.333050950018153,
"grad_norm": 0.10566890437995724,
"learning_rate": 7.178555740133694e-06,
"loss": 0.0437,
"num_input_tokens_seen": 907639040,
"step": 2410,
"train_runtime": 205007.1634,
"train_tokens_per_second": 4427.353
},
{
"epoch": 2.3378918068498122,
"grad_norm": 0.1078878711996976,
"learning_rate": 7.080092678287262e-06,
"loss": 0.0429,
"num_input_tokens_seen": 909548736,
"step": 2415,
"train_runtime": 205363.2901,
"train_tokens_per_second": 4428.974
},
{
"epoch": 2.3427326636814714,
"grad_norm": 0.10362678547369009,
"learning_rate": 6.982198046893912e-06,
"loss": 0.0403,
"num_input_tokens_seen": 911369184,
"step": 2420,
"train_runtime": 205601.0346,
"train_tokens_per_second": 4432.707
},
{
"epoch": 2.347573520513131,
"grad_norm": 0.1116147652899066,
"learning_rate": 6.884874951231407e-06,
"loss": 0.0444,
"num_input_tokens_seen": 913234896,
"step": 2425,
"train_runtime": 205896.2768,
"train_tokens_per_second": 4435.412
},
{
"epoch": 2.35241437734479,
"grad_norm": 0.11521220794776303,
"learning_rate": 6.788126478448046e-06,
"loss": 0.045,
"num_input_tokens_seen": 915148176,
"step": 2430,
"train_runtime": 206245.7823,
"train_tokens_per_second": 4437.173
},
{
"epoch": 2.357255234176449,
"grad_norm": 0.09281196214081752,
"learning_rate": 6.691955697464722e-06,
"loss": 0.0438,
"num_input_tokens_seen": 917092976,
"step": 2435,
"train_runtime": 206645.7743,
"train_tokens_per_second": 4437.995
},
{
"epoch": 2.3620960910081084,
"grad_norm": 0.100214030668592,
"learning_rate": 6.596365658877612e-06,
"loss": 0.0424,
"num_input_tokens_seen": 919066720,
"step": 2440,
"train_runtime": 207017.7612,
"train_tokens_per_second": 4439.555
},
{
"epoch": 2.3669369478397675,
"grad_norm": 0.10760036279429956,
"learning_rate": 6.501359394861378e-06,
"loss": 0.0417,
"num_input_tokens_seen": 920862256,
"step": 2445,
"train_runtime": 207284.7764,
"train_tokens_per_second": 4442.498
},
{
"epoch": 2.371777804671427,
"grad_norm": 0.10509350266248424,
"learning_rate": 6.4069399190729965e-06,
"loss": 0.045,
"num_input_tokens_seen": 922629344,
"step": 2450,
"train_runtime": 207527.687,
"train_tokens_per_second": 4445.813
},
{
"epoch": 2.376618661503086,
"grad_norm": 0.10714050286215131,
"learning_rate": 6.3131102265561595e-06,
"loss": 0.0428,
"num_input_tokens_seen": 924459968,
"step": 2455,
"train_runtime": 207811.3596,
"train_tokens_per_second": 4448.554
},
{
"epoch": 2.3814595183347453,
"grad_norm": 0.09927356415259088,
"learning_rate": 6.219873293646275e-06,
"loss": 0.0434,
"num_input_tokens_seen": 926282784,
"step": 2460,
"train_runtime": 208098.2228,
"train_tokens_per_second": 4451.181
},
{
"epoch": 2.3863003751664045,
"grad_norm": 0.10624426038475755,
"learning_rate": 6.127232077876044e-06,
"loss": 0.0423,
"num_input_tokens_seen": 928104928,
"step": 2465,
"train_runtime": 208396.5232,
"train_tokens_per_second": 4453.553
},
{
"epoch": 2.3911412319980636,
"grad_norm": 0.10065956898085184,
"learning_rate": 6.0351895178816675e-06,
"loss": 0.0414,
"num_input_tokens_seen": 930034976,
"step": 2470,
"train_runtime": 208753.2157,
"train_tokens_per_second": 4455.189
},
{
"epoch": 2.3959820888297227,
"grad_norm": 0.08919790357784825,
"learning_rate": 5.943748533309609e-06,
"loss": 0.0403,
"num_input_tokens_seen": 931869680,
"step": 2475,
"train_runtime": 209030.5157,
"train_tokens_per_second": 4458.056
},
{
"epoch": 2.400822945661382,
"grad_norm": 0.09926050843218681,
"learning_rate": 5.8529120247239945e-06,
"loss": 0.0426,
"num_input_tokens_seen": 933822272,
"step": 2480,
"train_runtime": 209411.1645,
"train_tokens_per_second": 4459.276
},
{
"epoch": 2.4056638024930415,
"grad_norm": 0.09286276749293895,
"learning_rate": 5.762682873514605e-06,
"loss": 0.041,
"num_input_tokens_seen": 935677664,
"step": 2485,
"train_runtime": 209681.396,
"train_tokens_per_second": 4462.378
},
{
"epoch": 2.4105046593247006,
"grad_norm": 0.11296254298586997,
"learning_rate": 5.673063941805451e-06,
"loss": 0.0445,
"num_input_tokens_seen": 937557696,
"step": 2490,
"train_runtime": 209997.2312,
"train_tokens_per_second": 4464.619
},
{
"epoch": 2.4153455161563597,
"grad_norm": 0.08754210776797587,
"learning_rate": 5.584058072364032e-06,
"loss": 0.0403,
"num_input_tokens_seen": 939447456,
"step": 2495,
"train_runtime": 210255.8593,
"train_tokens_per_second": 4468.115
},
{
"epoch": 2.420186372988019,
"grad_norm": 0.10698743548503842,
"learning_rate": 5.495668088511128e-06,
"loss": 0.0413,
"num_input_tokens_seen": 941372928,
"step": 2500,
"train_runtime": 210635.05,
"train_tokens_per_second": 4469.213
},
{
"epoch": 2.425027229819678,
"grad_norm": 0.11191046124395111,
"learning_rate": 5.407896794031245e-06,
"loss": 0.0435,
"num_input_tokens_seen": 943222032,
"step": 2505,
"train_runtime": 210947.9822,
"train_tokens_per_second": 4471.349
},
{
"epoch": 2.429868086651337,
"grad_norm": 0.11147955526705726,
"learning_rate": 5.3207469730836875e-06,
"loss": 0.0424,
"num_input_tokens_seen": 945141472,
"step": 2510,
"train_runtime": 211288.3155,
"train_tokens_per_second": 4473.231
},
{
"epoch": 2.4347089434829963,
"grad_norm": 0.11269801669995572,
"learning_rate": 5.2342213901142415e-06,
"loss": 0.042,
"num_input_tokens_seen": 946942384,
"step": 2515,
"train_runtime": 211559.8316,
"train_tokens_per_second": 4476.003
},
{
"epoch": 2.439549800314656,
"grad_norm": 0.10414158057001394,
"learning_rate": 5.1483227897674736e-06,
"loss": 0.0424,
"num_input_tokens_seen": 948797600,
"step": 2520,
"train_runtime": 211859.0494,
"train_tokens_per_second": 4478.438
},
{
"epoch": 2.444390657146315,
"grad_norm": 0.10225241274686434,
"learning_rate": 5.0630538967996824e-06,
"loss": 0.044,
"num_input_tokens_seen": 950738128,
"step": 2525,
"train_runtime": 212239.9612,
"train_tokens_per_second": 4479.543
},
{
"epoch": 2.449231513977974,
"grad_norm": 0.09486493161848354,
"learning_rate": 4.978417415992459e-06,
"loss": 0.0414,
"num_input_tokens_seen": 952683104,
"step": 2530,
"train_runtime": 212560.4677,
"train_tokens_per_second": 4481.939
},
{
"epoch": 2.4540723708096333,
"grad_norm": 0.10225202491694556,
"learning_rate": 4.8944160320668925e-06,
"loss": 0.044,
"num_input_tokens_seen": 954565248,
"step": 2535,
"train_runtime": 212864.2481,
"train_tokens_per_second": 4484.385
},
{
"epoch": 2.4589132276412924,
"grad_norm": 0.10555017743814563,
"learning_rate": 4.811052409598415e-06,
"loss": 0.0435,
"num_input_tokens_seen": 956458752,
"step": 2540,
"train_runtime": 213206.0267,
"train_tokens_per_second": 4486.077
},
{
"epoch": 2.4637540844729515,
"grad_norm": 0.10832795415710876,
"learning_rate": 4.728329192932277e-06,
"loss": 0.0434,
"num_input_tokens_seen": 958360512,
"step": 2545,
"train_runtime": 213541.3209,
"train_tokens_per_second": 4487.939
},
{
"epoch": 2.468594941304611,
"grad_norm": 0.10095120179330958,
"learning_rate": 4.64624900609964e-06,
"loss": 0.0433,
"num_input_tokens_seen": 960237872,
"step": 2550,
"train_runtime": 213826.9979,
"train_tokens_per_second": 4490.723
},
{
"epoch": 2.4734357981362702,
"grad_norm": 0.10772165766963647,
"learning_rate": 4.56481445273439e-06,
"loss": 0.0413,
"num_input_tokens_seen": 962247008,
"step": 2555,
"train_runtime": 214215.3446,
"train_tokens_per_second": 4491.961
},
{
"epoch": 2.4782766549679294,
"grad_norm": 0.09806092608971834,
"learning_rate": 4.4840281159905155e-06,
"loss": 0.0414,
"num_input_tokens_seen": 964170272,
"step": 2560,
"train_runtime": 214550.1729,
"train_tokens_per_second": 4493.915
},
{
"epoch": 2.4831175117995885,
"grad_norm": 0.09220012235718683,
"learning_rate": 4.403892558460177e-06,
"loss": 0.0412,
"num_input_tokens_seen": 966120000,
"step": 2565,
"train_runtime": 214929.4897,
"train_tokens_per_second": 4495.056
},
{
"epoch": 2.4879583686312476,
"grad_norm": 0.09514831510019058,
"learning_rate": 4.324410322092415e-06,
"loss": 0.0412,
"num_input_tokens_seen": 967960784,
"step": 2570,
"train_runtime": 215212.0348,
"train_tokens_per_second": 4497.708
},
{
"epoch": 2.4927992254629068,
"grad_norm": 0.08465241248211482,
"learning_rate": 4.2455839281125256e-06,
"loss": 0.0417,
"num_input_tokens_seen": 969844080,
"step": 2575,
"train_runtime": 215514.8865,
"train_tokens_per_second": 4500.126
},
{
"epoch": 2.4976400822945664,
"grad_norm": 0.09900416312304733,
"learning_rate": 4.167415876942085e-06,
"loss": 0.0417,
"num_input_tokens_seen": 971694160,
"step": 2580,
"train_runtime": 215818.3338,
"train_tokens_per_second": 4502.371
},
{
"epoch": 2.5024809391262255,
"grad_norm": 0.09456247404799463,
"learning_rate": 4.089908648119625e-06,
"loss": 0.0425,
"num_input_tokens_seen": 973545824,
"step": 2585,
"train_runtime": 216108.9629,
"train_tokens_per_second": 4504.884
},
{
"epoch": 2.5073217959578846,
"grad_norm": 0.1006524579480961,
"learning_rate": 4.013064700221991e-06,
"loss": 0.0411,
"num_input_tokens_seen": 975457776,
"step": 2590,
"train_runtime": 216449.4006,
"train_tokens_per_second": 4506.632
},
{
"epoch": 2.5121626527895438,
"grad_norm": 0.09861997049355122,
"learning_rate": 3.936886470786347e-06,
"loss": 0.0405,
"num_input_tokens_seen": 977366416,
"step": 2595,
"train_runtime": 216770.5711,
"train_tokens_per_second": 4508.76
},
{
"epoch": 2.517003509621203,
"grad_norm": 0.09885870437167427,
"learning_rate": 3.8613763762328695e-06,
"loss": 0.0432,
"num_input_tokens_seen": 979260288,
"step": 2600,
"train_runtime": 217072.79,
"train_tokens_per_second": 4511.207
},
{
"epoch": 2.521844366452862,
"grad_norm": 0.1044111303405238,
"learning_rate": 3.7865368117880688e-06,
"loss": 0.0441,
"num_input_tokens_seen": 981164192,
"step": 2605,
"train_runtime": 217397.5953,
"train_tokens_per_second": 4513.225
},
{
"epoch": 2.526685223284521,
"grad_norm": 0.09689580727353043,
"learning_rate": 3.712370151408842e-06,
"loss": 0.041,
"num_input_tokens_seen": 983019680,
"step": 2610,
"train_runtime": 217701.7754,
"train_tokens_per_second": 4515.442
},
{
"epoch": 2.5315260801161807,
"grad_norm": 0.07930860025277806,
"learning_rate": 3.6388787477071557e-06,
"loss": 0.0382,
"num_input_tokens_seen": 984946480,
"step": 2615,
"train_runtime": 218039.1448,
"train_tokens_per_second": 4517.292
},
{
"epoch": 2.53636693694784,
"grad_norm": 0.09701585395532773,
"learning_rate": 3.5660649318754153e-06,
"loss": 0.0431,
"num_input_tokens_seen": 986820704,
"step": 2620,
"train_runtime": 218364.0414,
"train_tokens_per_second": 4519.154
},
{
"epoch": 2.541207793779499,
"grad_norm": 0.08951915315467092,
"learning_rate": 3.493931013612528e-06,
"loss": 0.0404,
"num_input_tokens_seen": 988683712,
"step": 2625,
"train_runtime": 218662.1125,
"train_tokens_per_second": 4521.514
},
{
"epoch": 2.546048650611158,
"grad_norm": 0.10946918657325026,
"learning_rate": 3.4224792810506335e-06,
"loss": 0.0412,
"num_input_tokens_seen": 990613776,
"step": 2630,
"train_runtime": 219031.7427,
"train_tokens_per_second": 4522.695
},
{
"epoch": 2.5508895074428173,
"grad_norm": 0.09243649296411362,
"learning_rate": 3.3517120006825137e-06,
"loss": 0.0421,
"num_input_tokens_seen": 992516752,
"step": 2635,
"train_runtime": 219368.5843,
"train_tokens_per_second": 4524.425
},
{
"epoch": 2.5557303642744764,
"grad_norm": 0.11072560597436613,
"learning_rate": 3.2816314172897183e-06,
"loss": 0.0433,
"num_input_tokens_seen": 994379968,
"step": 2640,
"train_runtime": 219685.0843,
"train_tokens_per_second": 4526.388
},
{
"epoch": 2.5605712211061356,
"grad_norm": 0.1058803495743547,
"learning_rate": 3.2122397538713432e-06,
"loss": 0.0413,
"num_input_tokens_seen": 996372960,
"step": 2645,
"train_runtime": 220112.659,
"train_tokens_per_second": 4526.65
},
{
"epoch": 2.565412077937795,
"grad_norm": 0.09586921377974332,
"learning_rate": 3.143539211573518e-06,
"loss": 0.0445,
"num_input_tokens_seen": 998147504,
"step": 2650,
"train_runtime": 220369.4226,
"train_tokens_per_second": 4529.428
},
{
"epoch": 2.5702529347694543,
"grad_norm": 0.100735105625526,
"learning_rate": 3.075531969619594e-06,
"loss": 0.0402,
"num_input_tokens_seen": 1000010032,
"step": 2655,
"train_runtime": 220683.995,
"train_tokens_per_second": 4531.412
},
{
"epoch": 2.5750937916011134,
"grad_norm": 0.08157917824449015,
"learning_rate": 3.00822018524099e-06,
"loss": 0.0394,
"num_input_tokens_seen": 1001911504,
"step": 2660,
"train_runtime": 220992.259,
"train_tokens_per_second": 4533.695
},
{
"epoch": 2.5799346484327725,
"grad_norm": 0.10671166825123399,
"learning_rate": 2.9416059936088176e-06,
"loss": 0.0405,
"num_input_tokens_seen": 1003915536,
"step": 2665,
"train_runtime": 221370.367,
"train_tokens_per_second": 4535.004
},
{
"epoch": 2.5847755052644317,
"grad_norm": 0.10485178372092739,
"learning_rate": 2.8756915077660816e-06,
"loss": 0.042,
"num_input_tokens_seen": 1005786320,
"step": 2670,
"train_runtime": 221680.2811,
"train_tokens_per_second": 4537.103
},
{
"epoch": 2.5896163620960913,
"grad_norm": 0.11239886407003709,
"learning_rate": 2.810478818560719e-06,
"loss": 0.046,
"num_input_tokens_seen": 1007566272,
"step": 2675,
"train_runtime": 221965.9319,
"train_tokens_per_second": 4539.283
},
{
"epoch": 2.59445721892775,
"grad_norm": 0.10703312505536274,
"learning_rate": 2.745969994579231e-06,
"loss": 0.0429,
"num_input_tokens_seen": 1009450528,
"step": 2680,
"train_runtime": 222294.4882,
"train_tokens_per_second": 4541.051
},
{
"epoch": 2.5992980757594095,
"grad_norm": 0.09294201066454781,
"learning_rate": 2.6821670820810883e-06,
"loss": 0.0413,
"num_input_tokens_seen": 1011300880,
"step": 2685,
"train_runtime": 222609.6014,
"train_tokens_per_second": 4542.935
},
{
"epoch": 2.6041389325910687,
"grad_norm": 0.09973495575855759,
"learning_rate": 2.6190721049338145e-06,
"loss": 0.0406,
"num_input_tokens_seen": 1013140688,
"step": 2690,
"train_runtime": 222887.9,
"train_tokens_per_second": 4545.517
},
{
"epoch": 2.608979789422728,
"grad_norm": 0.100229670298609,
"learning_rate": 2.5566870645487906e-06,
"loss": 0.0423,
"num_input_tokens_seen": 1015019296,
"step": 2695,
"train_runtime": 223209.5025,
"train_tokens_per_second": 4547.384
},
{
"epoch": 2.613820646254387,
"grad_norm": 0.08644793584438773,
"learning_rate": 2.495013939817767e-06,
"loss": 0.0395,
"num_input_tokens_seen": 1016892176,
"step": 2700,
"train_runtime": 223525.4721,
"train_tokens_per_second": 4549.335
},
{
"epoch": 2.618661503086046,
"grad_norm": 0.09650950829215955,
"learning_rate": 2.434054687050091e-06,
"loss": 0.0405,
"num_input_tokens_seen": 1018797600,
"step": 2705,
"train_runtime": 223827.1085,
"train_tokens_per_second": 4551.717
},
{
"epoch": 2.6235023599177056,
"grad_norm": 0.12423581499512551,
"learning_rate": 2.3738112399106644e-06,
"loss": 0.0454,
"num_input_tokens_seen": 1020718416,
"step": 2710,
"train_runtime": 224190.0696,
"train_tokens_per_second": 4552.915
},
{
"epoch": 2.6283432167493648,
"grad_norm": 0.10989337129875472,
"learning_rate": 2.314285509358566e-06,
"loss": 0.0428,
"num_input_tokens_seen": 1022524336,
"step": 2715,
"train_runtime": 224455.0174,
"train_tokens_per_second": 4555.587
},
{
"epoch": 2.633184073581024,
"grad_norm": 0.09039073713194058,
"learning_rate": 2.255479383586509e-06,
"loss": 0.0406,
"num_input_tokens_seen": 1024388192,
"step": 2720,
"train_runtime": 224743.9654,
"train_tokens_per_second": 4558.023
},
{
"epoch": 2.638024930412683,
"grad_norm": 0.08953330757222618,
"learning_rate": 2.197394727960861e-06,
"loss": 0.0408,
"num_input_tokens_seen": 1026219264,
"step": 2725,
"train_runtime": 225055.269,
"train_tokens_per_second": 4559.854
},
{
"epoch": 2.642865787244342,
"grad_norm": 0.08090268583905104,
"learning_rate": 2.1400333849625338e-06,
"loss": 0.0392,
"num_input_tokens_seen": 1028260624,
"step": 2730,
"train_runtime": 225415.8787,
"train_tokens_per_second": 4561.616
},
{
"epoch": 2.6477066440760013,
"grad_norm": 0.10367619160660342,
"learning_rate": 2.083397174128518e-06,
"loss": 0.0415,
"num_input_tokens_seen": 1030159584,
"step": 2735,
"train_runtime": 225741.8649,
"train_tokens_per_second": 4563.441
},
{
"epoch": 2.6525475009076604,
"grad_norm": 0.11074351346114625,
"learning_rate": 2.027487891994162e-06,
"loss": 0.0432,
"num_input_tokens_seen": 1032081920,
"step": 2740,
"train_runtime": 226092.6663,
"train_tokens_per_second": 4564.862
},
{
"epoch": 2.65738835773932,
"grad_norm": 0.10039057936660997,
"learning_rate": 1.9723073120361924e-06,
"loss": 0.0425,
"num_input_tokens_seen": 1033931904,
"step": 2745,
"train_runtime": 226382.3218,
"train_tokens_per_second": 4567.194
},
{
"epoch": 2.662229214570979,
"grad_norm": 0.09363138188924082,
"learning_rate": 1.9178571846164532e-06,
"loss": 0.0397,
"num_input_tokens_seen": 1035806768,
"step": 2750,
"train_runtime": 226704.6785,
"train_tokens_per_second": 4568.97
},
{
"epoch": 2.6670700714026383,
"grad_norm": 0.08714334312836239,
"learning_rate": 1.8641392369263933e-06,
"loss": 0.04,
"num_input_tokens_seen": 1037779856,
"step": 2755,
"train_runtime": 227072.2692,
"train_tokens_per_second": 4570.262
},
{
"epoch": 2.6719109282342974,
"grad_norm": 0.10402712457251752,
"learning_rate": 1.8111551729322662e-06,
"loss": 0.0439,
"num_input_tokens_seen": 1039606192,
"step": 2760,
"train_runtime": 227397.5471,
"train_tokens_per_second": 4571.756
},
{
"epoch": 2.6767517850659566,
"grad_norm": 0.10242769460180877,
"learning_rate": 1.7589066733210814e-06,
"loss": 0.0473,
"num_input_tokens_seen": 1041352016,
"step": 2765,
"train_runtime": 227680.2847,
"train_tokens_per_second": 4573.747
},
{
"epoch": 2.681592641897616,
"grad_norm": 0.10096005023224232,
"learning_rate": 1.7073953954472949e-06,
"loss": 0.0432,
"num_input_tokens_seen": 1043209584,
"step": 2770,
"train_runtime": 228004.7054,
"train_tokens_per_second": 4575.386
},
{
"epoch": 2.686433498729275,
"grad_norm": 0.10430771290514242,
"learning_rate": 1.6566229732802501e-06,
"loss": 0.0423,
"num_input_tokens_seen": 1045070400,
"step": 2775,
"train_runtime": 228328.0951,
"train_tokens_per_second": 4577.056
},
{
"epoch": 2.6912743555609344,
"grad_norm": 0.09976539169225217,
"learning_rate": 1.6065910173523101e-06,
"loss": 0.0404,
"num_input_tokens_seen": 1047027216,
"step": 2780,
"train_runtime": 228649.8742,
"train_tokens_per_second": 4579.173
},
{
"epoch": 2.6961152123925936,
"grad_norm": 0.09965543945612194,
"learning_rate": 1.5573011147078236e-06,
"loss": 0.0407,
"num_input_tokens_seen": 1048933616,
"step": 2785,
"train_runtime": 228968.4155,
"train_tokens_per_second": 4581.128
},
{
"epoch": 2.7009560692242527,
"grad_norm": 0.09982142335711996,
"learning_rate": 1.5087548288527291e-06,
"loss": 0.0416,
"num_input_tokens_seen": 1050730672,
"step": 2790,
"train_runtime": 229238.8178,
"train_tokens_per_second": 4583.563
},
{
"epoch": 2.705796926055912,
"grad_norm": 0.10720846842326023,
"learning_rate": 1.4609536997049977e-06,
"loss": 0.0415,
"num_input_tokens_seen": 1052607296,
"step": 2795,
"train_runtime": 229546.0728,
"train_tokens_per_second": 4585.604
},
{
"epoch": 2.710637782887571,
"grad_norm": 0.1048290688754556,
"learning_rate": 1.4138992435457688e-06,
"loss": 0.041,
"num_input_tokens_seen": 1054530832,
"step": 2800,
"train_runtime": 229875.6229,
"train_tokens_per_second": 4587.397
},
{
"epoch": 2.7154786397192305,
"grad_norm": 0.10487087810225432,
"learning_rate": 1.3675929529712555e-06,
"loss": 0.0404,
"num_input_tokens_seen": 1056509856,
"step": 2805,
"train_runtime": 230229.7535,
"train_tokens_per_second": 4588.937
},
{
"epoch": 2.7203194965508897,
"grad_norm": 0.10984325530051703,
"learning_rate": 1.3220362968454026e-06,
"loss": 0.041,
"num_input_tokens_seen": 1058388320,
"step": 2810,
"train_runtime": 230583.9386,
"train_tokens_per_second": 4590.035
},
{
"epoch": 2.725160353382549,
"grad_norm": 0.09575719645821536,
"learning_rate": 1.277230720253289e-06,
"loss": 0.0416,
"num_input_tokens_seen": 1060208912,
"step": 2815,
"train_runtime": 230862.7987,
"train_tokens_per_second": 4592.377
},
{
"epoch": 2.730001210214208,
"grad_norm": 0.0952849105139362,
"learning_rate": 1.2331776444552939e-06,
"loss": 0.042,
"num_input_tokens_seen": 1062178720,
"step": 2820,
"train_runtime": 231234.4358,
"train_tokens_per_second": 4593.514
},
{
"epoch": 2.734842067045867,
"grad_norm": 0.09720163766984548,
"learning_rate": 1.1898784668419927e-06,
"loss": 0.0409,
"num_input_tokens_seen": 1064115760,
"step": 2825,
"train_runtime": 231566.8629,
"train_tokens_per_second": 4595.285
},
{
"epoch": 2.739682923877526,
"grad_norm": 0.10713711730304852,
"learning_rate": 1.1473345608898789e-06,
"loss": 0.0433,
"num_input_tokens_seen": 1065971408,
"step": 2830,
"train_runtime": 231877.52,
"train_tokens_per_second": 4597.131
},
{
"epoch": 2.7445237807091853,
"grad_norm": 0.09631173848061805,
"learning_rate": 1.10554727611773e-06,
"loss": 0.0433,
"num_input_tokens_seen": 1067891024,
"step": 2835,
"train_runtime": 232238.3471,
"train_tokens_per_second": 4598.254
},
{
"epoch": 2.749364637540845,
"grad_norm": 0.10739207946404603,
"learning_rate": 1.0645179380438657e-06,
"loss": 0.0433,
"num_input_tokens_seen": 1069744096,
"step": 2840,
"train_runtime": 232579.3058,
"train_tokens_per_second": 4599.481
},
{
"epoch": 2.754205494372504,
"grad_norm": 0.10233938837342275,
"learning_rate": 1.0242478481440498e-06,
"loss": 0.0395,
"num_input_tokens_seen": 1071601856,
"step": 2845,
"train_runtime": 232856.6503,
"train_tokens_per_second": 4601.981
},
{
"epoch": 2.759046351204163,
"grad_norm": 0.09778768115212016,
"learning_rate": 9.847382838102492e-07,
"loss": 0.0421,
"num_input_tokens_seen": 1073480480,
"step": 2850,
"train_runtime": 233167.1054,
"train_tokens_per_second": 4603.91
},
{
"epoch": 2.7638872080358223,
"grad_norm": 0.08134168754707406,
"learning_rate": 9.459904983100704e-07,
"loss": 0.0404,
"num_input_tokens_seen": 1075401632,
"step": 2855,
"train_runtime": 233474.4543,
"train_tokens_per_second": 4606.078
},
{
"epoch": 2.7687280648674815,
"grad_norm": 0.10586368696758197,
"learning_rate": 9.080057207470405e-07,
"loss": 0.0379,
"num_input_tokens_seen": 1077335872,
"step": 2860,
"train_runtime": 233807.1971,
"train_tokens_per_second": 4607.796
},
{
"epoch": 2.7735689216991406,
"grad_norm": 0.09864439204162297,
"learning_rate": 8.707851560216112e-07,
"loss": 0.0403,
"num_input_tokens_seen": 1079310576,
"step": 2865,
"train_runtime": 234165.3118,
"train_tokens_per_second": 4609.182
},
{
"epoch": 2.7784097785307997,
"grad_norm": 0.09763664283200235,
"learning_rate": 8.343299847929226e-07,
"loss": 0.0387,
"num_input_tokens_seen": 1081245280,
"step": 2870,
"train_runtime": 234484.5006,
"train_tokens_per_second": 4611.159
},
{
"epoch": 2.7832506353624593,
"grad_norm": 0.10339662193442319,
"learning_rate": 7.986413634413686e-07,
"loss": 0.0423,
"num_input_tokens_seen": 1083177952,
"step": 2875,
"train_runtime": 234868.9456,
"train_tokens_per_second": 4611.84
},
{
"epoch": 2.7880914921941184,
"grad_norm": 0.10893179032282667,
"learning_rate": 7.637204240319163e-07,
"loss": 0.0415,
"num_input_tokens_seen": 1085074640,
"step": 2880,
"train_runtime": 235198.1723,
"train_tokens_per_second": 4613.448
},
{
"epoch": 2.7929323490257776,
"grad_norm": 0.09451603653033806,
"learning_rate": 7.295682742781862e-07,
"loss": 0.04,
"num_input_tokens_seen": 1086903600,
"step": 2885,
"train_runtime": 235485.5269,
"train_tokens_per_second": 4615.586
},
{
"epoch": 2.7977732058574367,
"grad_norm": 0.09247794261791085,
"learning_rate": 6.961859975073121e-07,
"loss": 0.0407,
"num_input_tokens_seen": 1088720960,
"step": 2890,
"train_runtime": 235741.8558,
"train_tokens_per_second": 4618.276
},
{
"epoch": 2.802614062689096,
"grad_norm": 0.09819027331260509,
"learning_rate": 6.635746526255981e-07,
"loss": 0.0399,
"num_input_tokens_seen": 1090542208,
"step": 2895,
"train_runtime": 236021.1683,
"train_tokens_per_second": 4620.527
},
{
"epoch": 2.8074549195207554,
"grad_norm": 0.10508939236953255,
"learning_rate": 6.317352740849048e-07,
"loss": 0.0407,
"num_input_tokens_seen": 1092439568,
"step": 2900,
"train_runtime": 236348.1264,
"train_tokens_per_second": 4622.163
},
{
"epoch": 2.812295776352414,
"grad_norm": 0.1127436229338298,
"learning_rate": 6.006688718498549e-07,
"loss": 0.0425,
"num_input_tokens_seen": 1094246688,
"step": 2905,
"train_runtime": 236640.2834,
"train_tokens_per_second": 4624.093
},
{
"epoch": 2.8171366331840737,
"grad_norm": 0.09425104412787287,
"learning_rate": 5.703764313657795e-07,
"loss": 0.0426,
"num_input_tokens_seen": 1096143712,
"step": 2910,
"train_runtime": 236969.0877,
"train_tokens_per_second": 4625.682
},
{
"epoch": 2.821977490015733,
"grad_norm": 0.0964656395201385,
"learning_rate": 5.408589135274755e-07,
"loss": 0.0411,
"num_input_tokens_seen": 1098111104,
"step": 2915,
"train_runtime": 237285.2741,
"train_tokens_per_second": 4627.81
},
{
"epoch": 2.826818346847392,
"grad_norm": 0.10779087274875156,
"learning_rate": 5.121172546487196e-07,
"loss": 0.0407,
"num_input_tokens_seen": 1100058224,
"step": 2920,
"train_runtime": 237651.5376,
"train_tokens_per_second": 4628.871
},
{
"epoch": 2.831659203679051,
"grad_norm": 0.10847631726351929,
"learning_rate": 4.841523664325581e-07,
"loss": 0.0397,
"num_input_tokens_seen": 1101896720,
"step": 2925,
"train_runtime": 237940.6724,
"train_tokens_per_second": 4630.973
},
{
"epoch": 2.8365000605107102,
"grad_norm": 0.09624461904411019,
"learning_rate": 4.5696513594240264e-07,
"loss": 0.0412,
"num_input_tokens_seen": 1103744592,
"step": 2930,
"train_runtime": 238212.5813,
"train_tokens_per_second": 4633.444
},
{
"epoch": 2.84134091734237,
"grad_norm": 0.1121598440101494,
"learning_rate": 4.305564255738831e-07,
"loss": 0.0407,
"num_input_tokens_seen": 1105651136,
"step": 2935,
"train_runtime": 238531.5307,
"train_tokens_per_second": 4635.241
},
{
"epoch": 2.846181774174029,
"grad_norm": 0.10501345290205552,
"learning_rate": 4.0492707302749176e-07,
"loss": 0.0412,
"num_input_tokens_seen": 1107467904,
"step": 2940,
"train_runtime": 238807.7005,
"train_tokens_per_second": 4637.488
},
{
"epoch": 2.851022631005688,
"grad_norm": 0.09913381095639745,
"learning_rate": 3.800778912820102e-07,
"loss": 0.041,
"num_input_tokens_seen": 1109281136,
"step": 2945,
"train_runtime": 239102.1626,
"train_tokens_per_second": 4639.361
},
{
"epoch": 2.8558634878373472,
"grad_norm": 0.10690920603697875,
"learning_rate": 3.560096685687325e-07,
"loss": 0.041,
"num_input_tokens_seen": 1111166496,
"step": 2950,
"train_runtime": 239424.9583,
"train_tokens_per_second": 4640.98
},
{
"epoch": 2.8607043446690064,
"grad_norm": 0.10371511929667834,
"learning_rate": 3.32723168346441e-07,
"loss": 0.0408,
"num_input_tokens_seen": 1113024464,
"step": 2955,
"train_runtime": 239736.7802,
"train_tokens_per_second": 4642.694
},
{
"epoch": 2.8655452015006655,
"grad_norm": 0.09279356116922857,
"learning_rate": 3.102191292772144e-07,
"loss": 0.0428,
"num_input_tokens_seen": 1114789808,
"step": 2960,
"train_runtime": 239990.4972,
"train_tokens_per_second": 4645.141
},
{
"epoch": 2.8703860583323246,
"grad_norm": 0.10484566419466071,
"learning_rate": 2.884982652029716e-07,
"loss": 0.0414,
"num_input_tokens_seen": 1116715504,
"step": 2965,
"train_runtime": 240306.4348,
"train_tokens_per_second": 4647.048
},
{
"epoch": 2.875226915163984,
"grad_norm": 0.09371900146099076,
"learning_rate": 2.6756126512285094e-07,
"loss": 0.0423,
"num_input_tokens_seen": 1118661392,
"step": 2970,
"train_runtime": 240658.0326,
"train_tokens_per_second": 4648.344
},
{
"epoch": 2.8800677719956433,
"grad_norm": 0.08968284205417433,
"learning_rate": 2.4740879317133314e-07,
"loss": 0.043,
"num_input_tokens_seen": 1120523488,
"step": 2975,
"train_runtime": 240990.9356,
"train_tokens_per_second": 4649.65
},
{
"epoch": 2.8849086288273025,
"grad_norm": 0.1194432143401207,
"learning_rate": 2.2804148859719433e-07,
"loss": 0.0425,
"num_input_tokens_seen": 1122329216,
"step": 2980,
"train_runtime": 241267.0534,
"train_tokens_per_second": 4651.813
},
{
"epoch": 2.8897494856589616,
"grad_norm": 0.10332352909612158,
"learning_rate": 2.0945996574321392e-07,
"loss": 0.0423,
"num_input_tokens_seen": 1124108576,
"step": 2985,
"train_runtime": 241512.3485,
"train_tokens_per_second": 4654.456
},
{
"epoch": 2.8945903424906207,
"grad_norm": 0.09902903701655406,
"learning_rate": 1.9166481402669856e-07,
"loss": 0.0416,
"num_input_tokens_seen": 1125941808,
"step": 2990,
"train_runtime": 241794.3455,
"train_tokens_per_second": 4656.609
},
{
"epoch": 2.89943119932228,
"grad_norm": 0.10194619378485503,
"learning_rate": 1.7465659792077484e-07,
"loss": 0.0412,
"num_input_tokens_seen": 1127752624,
"step": 2995,
"train_runtime": 242080.3868,
"train_tokens_per_second": 4658.587
},
{
"epoch": 2.904272056153939,
"grad_norm": 0.11088193358770991,
"learning_rate": 1.5843585693648967e-07,
"loss": 0.0418,
"num_input_tokens_seen": 1129601232,
"step": 3000,
"train_runtime": 242368.9222,
"train_tokens_per_second": 4660.669
},
{
"epoch": 2.9091129129855986,
"grad_norm": 0.09014663812275092,
"learning_rate": 1.4300310560570184e-07,
"loss": 0.041,
"num_input_tokens_seen": 1131446736,
"step": 3005,
"train_runtime": 242641.5856,
"train_tokens_per_second": 4663.037
},
{
"epoch": 2.9139537698172577,
"grad_norm": 0.10599016609672014,
"learning_rate": 1.2835883346474786e-07,
"loss": 0.0402,
"num_input_tokens_seen": 1133359440,
"step": 3010,
"train_runtime": 242967.2782,
"train_tokens_per_second": 4664.659
},
{
"epoch": 2.918794626648917,
"grad_norm": 0.09666954905813178,
"learning_rate": 1.1450350503892648e-07,
"loss": 0.0413,
"num_input_tokens_seen": 1135150272,
"step": 3015,
"train_runtime": 243234.1704,
"train_tokens_per_second": 4666.903
},
{
"epoch": 2.923635483480576,
"grad_norm": 0.0952319372122157,
"learning_rate": 1.014375598277495e-07,
"loss": 0.0397,
"num_input_tokens_seen": 1137058240,
"step": 3020,
"train_runtime": 243553.8042,
"train_tokens_per_second": 4668.612
},
{
"epoch": 2.928476340312235,
"grad_norm": 0.0961571420112863,
"learning_rate": 8.916141229101671e-08,
"loss": 0.0423,
"num_input_tokens_seen": 1138880976,
"step": 3025,
"train_runtime": 243833.0558,
"train_tokens_per_second": 4670.741
},
{
"epoch": 2.9333171971438947,
"grad_norm": 0.10748828390119694,
"learning_rate": 7.767545183565983e-08,
"loss": 0.0417,
"num_input_tokens_seen": 1140802272,
"step": 3030,
"train_runtime": 244185.0486,
"train_tokens_per_second": 4671.876
},
{
"epoch": 2.9381580539755534,
"grad_norm": 0.09453143585948885,
"learning_rate": 6.69800428033912e-08,
"loss": 0.0408,
"num_input_tokens_seen": 1142726336,
"step": 3035,
"train_runtime": 244534.0926,
"train_tokens_per_second": 4673.076
},
{
"epoch": 2.942998910807213,
"grad_norm": 0.10758820416109051,
"learning_rate": 5.707552445914366e-08,
"loss": 0.041,
"num_input_tokens_seen": 1144590976,
"step": 3040,
"train_runtime": 244862.1006,
"train_tokens_per_second": 4674.431
},
{
"epoch": 2.947839767638872,
"grad_norm": 0.09221176210129879,
"learning_rate": 4.7962210980317946e-08,
"loss": 0.0408,
"num_input_tokens_seen": 1146496640,
"step": 3045,
"train_runtime": 245189.2651,
"train_tokens_per_second": 4675.966
},
{
"epoch": 2.9526806244705313,
"grad_norm": 0.10331440557683293,
"learning_rate": 3.964039144680465e-08,
"loss": 0.0418,
"num_input_tokens_seen": 1148333472,
"step": 3050,
"train_runtime": 245482.1925,
"train_tokens_per_second": 4677.869
},
{
"epoch": 2.9575214813021904,
"grad_norm": 0.08747696767377501,
"learning_rate": 3.2110329831824825e-08,
"loss": 0.0423,
"num_input_tokens_seen": 1150262528,
"step": 3055,
"train_runtime": 245807.9806,
"train_tokens_per_second": 4679.517
},
{
"epoch": 2.9623623381338495,
"grad_norm": 0.08985631847344186,
"learning_rate": 2.537226499355616e-08,
"loss": 0.0384,
"num_input_tokens_seen": 1152112096,
"step": 3060,
"train_runtime": 246091.225,
"train_tokens_per_second": 4681.646
},
{
"epoch": 2.967203194965509,
"grad_norm": 0.10286034088749602,
"learning_rate": 1.942641066754458e-08,
"loss": 0.0419,
"num_input_tokens_seen": 1153978224,
"step": 3065,
"train_runtime": 246407.1175,
"train_tokens_per_second": 4683.218
},
{
"epoch": 2.9720440517971682,
"grad_norm": 0.09984385756001556,
"learning_rate": 1.4272955459937453e-08,
"loss": 0.0414,
"num_input_tokens_seen": 1155885360,
"step": 3070,
"train_runtime": 246753.738,
"train_tokens_per_second": 4684.368
},
{
"epoch": 2.9768849086288274,
"grad_norm": 0.09194901699543462,
"learning_rate": 9.912062841496705e-09,
"loss": 0.0387,
"num_input_tokens_seen": 1157755504,
"step": 3075,
"train_runtime": 247066.5954,
"train_tokens_per_second": 4686.006
},
{
"epoch": 2.9817257654604865,
"grad_norm": 0.10514662226999139,
"learning_rate": 6.343871142411306e-09,
"loss": 0.042,
"num_input_tokens_seen": 1159608192,
"step": 3080,
"train_runtime": 247360.5627,
"train_tokens_per_second": 4687.927
},
{
"epoch": 2.9865666222921456,
"grad_norm": 0.09518313021004897,
"learning_rate": 3.568493547909113e-09,
"loss": 0.0407,
"num_input_tokens_seen": 1161579776,
"step": 3085,
"train_runtime": 247715.5527,
"train_tokens_per_second": 4689.168
},
{
"epoch": 2.991407479123805,
"grad_norm": 0.09017165474805426,
"learning_rate": 1.586018094670849e-09,
"loss": 0.0391,
"num_input_tokens_seen": 1163408592,
"step": 3090,
"train_runtime": 247983.5908,
"train_tokens_per_second": 4691.474
},
{
"epoch": 2.996248335955464,
"grad_norm": 0.1022357668004034,
"learning_rate": 3.965076680351176e-10,
"loss": 0.0417,
"num_input_tokens_seen": 1165284400,
"step": 3095,
"train_runtime": 248298.2933,
"train_tokens_per_second": 4693.083
},
{
"epoch": 3.0,
"num_input_tokens_seen": 1166793552,
"step": 3099,
"total_flos": 5942698242998272.0,
"train_loss": 0.21310824125995556,
"train_runtime": 248614.6933,
"train_samples_per_second": 1.595,
"train_steps_per_second": 0.012
}
],
"logging_steps": 5,
"max_steps": 3099,
"num_input_tokens_seen": 1166793552,
"num_train_epochs": 3,
"save_steps": 7000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5942698242998272.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}