{ "best_metric": 0.750106155872345, "best_model_checkpoint": "sean_test_out_large/checkpoint-108334", "epoch": 1.0, "eval_steps": 500, "global_step": 108334, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 1387778.25, "learning_rate": 5.972307862720845e-05, "loss": 1.1897, "step": 500 }, { "epoch": 0.01, "grad_norm": 664302.625, "learning_rate": 5.94461572544169e-05, "loss": 1.0759, "step": 1000 }, { "epoch": 0.01, "grad_norm": 428486.84375, "learning_rate": 5.9169235881625344e-05, "loss": 1.0728, "step": 1500 }, { "epoch": 0.02, "grad_norm": 299683.625, "learning_rate": 5.8892314508833796e-05, "loss": 1.0409, "step": 2000 }, { "epoch": 0.02, "grad_norm": 503517.40625, "learning_rate": 5.861539313604224e-05, "loss": 1.0335, "step": 2500 }, { "epoch": 0.03, "grad_norm": 657503.8125, "learning_rate": 5.833847176325069e-05, "loss": 0.9533, "step": 3000 }, { "epoch": 0.03, "grad_norm": 533787.3125, "learning_rate": 5.806155039045914e-05, "loss": 0.9992, "step": 3500 }, { "epoch": 0.04, "grad_norm": 913784.6875, "learning_rate": 5.7784629017667584e-05, "loss": 0.9657, "step": 4000 }, { "epoch": 0.04, "grad_norm": 310152.21875, "learning_rate": 5.7507707644876036e-05, "loss": 0.9652, "step": 4500 }, { "epoch": 0.05, "grad_norm": 309690.34375, "learning_rate": 5.723078627208448e-05, "loss": 0.969, "step": 5000 }, { "epoch": 0.05, "grad_norm": 211456.5, "learning_rate": 5.695386489929293e-05, "loss": 0.9515, "step": 5500 }, { "epoch": 0.06, "grad_norm": 391416.65625, "learning_rate": 5.667694352650138e-05, "loss": 0.9221, "step": 6000 }, { "epoch": 0.06, "grad_norm": 221054.703125, "learning_rate": 5.6400022153709824e-05, "loss": 0.9679, "step": 6500 }, { "epoch": 0.06, "grad_norm": 472076.15625, "learning_rate": 5.612310078091827e-05, "loss": 0.9519, "step": 7000 }, { "epoch": 0.07, "grad_norm": 988913.375, "learning_rate": 5.584617940812672e-05, "loss": 0.9447, "step": 7500 }, { "epoch": 0.07, "grad_norm": 661127.75, "learning_rate": 5.556925803533517e-05, "loss": 0.9009, "step": 8000 }, { "epoch": 0.08, "grad_norm": 212504.640625, "learning_rate": 5.529233666254362e-05, "loss": 0.9456, "step": 8500 }, { "epoch": 0.08, "grad_norm": 473405.96875, "learning_rate": 5.5015415289752064e-05, "loss": 0.9385, "step": 9000 }, { "epoch": 0.09, "grad_norm": 266084.78125, "learning_rate": 5.473849391696051e-05, "loss": 1.0131, "step": 9500 }, { "epoch": 0.09, "grad_norm": 491548.53125, "learning_rate": 5.446157254416896e-05, "loss": 0.9254, "step": 10000 }, { "epoch": 0.1, "grad_norm": 267938.4375, "learning_rate": 5.418465117137741e-05, "loss": 0.9101, "step": 10500 }, { "epoch": 0.1, "grad_norm": 258540.84375, "learning_rate": 5.390772979858586e-05, "loss": 0.9234, "step": 11000 }, { "epoch": 0.11, "grad_norm": 200250.609375, "learning_rate": 5.3630808425794305e-05, "loss": 0.9157, "step": 11500 }, { "epoch": 0.11, "grad_norm": 480926.25, "learning_rate": 5.335388705300275e-05, "loss": 0.9532, "step": 12000 }, { "epoch": 0.12, "grad_norm": 593979.4375, "learning_rate": 5.30769656802112e-05, "loss": 0.9184, "step": 12500 }, { "epoch": 0.12, "grad_norm": 686144.1875, "learning_rate": 5.280004430741965e-05, "loss": 0.9215, "step": 13000 }, { "epoch": 0.12, "grad_norm": 582782.0, "learning_rate": 5.25231229346281e-05, "loss": 0.9151, "step": 13500 }, { "epoch": 0.13, "grad_norm": 554729.875, "learning_rate": 5.2246201561836545e-05, "loss": 0.9042, "step": 14000 }, { "epoch": 0.13, "grad_norm": 423190.0, "learning_rate": 5.196928018904499e-05, "loss": 0.9114, "step": 14500 }, { "epoch": 0.14, "grad_norm": 724497.5, "learning_rate": 5.169235881625344e-05, "loss": 0.9123, "step": 15000 }, { "epoch": 0.14, "grad_norm": 333863.78125, "learning_rate": 5.141543744346189e-05, "loss": 0.9237, "step": 15500 }, { "epoch": 0.15, "grad_norm": 282522.5, "learning_rate": 5.113851607067033e-05, "loss": 0.8927, "step": 16000 }, { "epoch": 0.15, "grad_norm": 326572.34375, "learning_rate": 5.0861594697878785e-05, "loss": 0.9256, "step": 16500 }, { "epoch": 0.16, "grad_norm": 454635.34375, "learning_rate": 5.058467332508723e-05, "loss": 0.8991, "step": 17000 }, { "epoch": 0.16, "grad_norm": 306963.25, "learning_rate": 5.030775195229568e-05, "loss": 0.8937, "step": 17500 }, { "epoch": 0.17, "grad_norm": 580224.0, "learning_rate": 5.003083057950413e-05, "loss": 0.9108, "step": 18000 }, { "epoch": 0.17, "grad_norm": 549220.75, "learning_rate": 4.975390920671257e-05, "loss": 0.9307, "step": 18500 }, { "epoch": 0.18, "grad_norm": 188370.0, "learning_rate": 4.9476987833921025e-05, "loss": 0.9009, "step": 19000 }, { "epoch": 0.18, "grad_norm": 535151.8125, "learning_rate": 4.920006646112947e-05, "loss": 0.8773, "step": 19500 }, { "epoch": 0.18, "grad_norm": 675718.1875, "learning_rate": 4.892314508833792e-05, "loss": 0.8984, "step": 20000 }, { "epoch": 0.19, "grad_norm": 247562.0, "learning_rate": 4.864622371554637e-05, "loss": 0.8896, "step": 20500 }, { "epoch": 0.19, "grad_norm": 412286.65625, "learning_rate": 4.836930234275481e-05, "loss": 0.8914, "step": 21000 }, { "epoch": 0.2, "grad_norm": 362339.53125, "learning_rate": 4.8092380969963265e-05, "loss": 0.8906, "step": 21500 }, { "epoch": 0.2, "grad_norm": 307259.8125, "learning_rate": 4.781545959717171e-05, "loss": 0.8615, "step": 22000 }, { "epoch": 0.21, "grad_norm": 264488.625, "learning_rate": 4.753853822438016e-05, "loss": 0.896, "step": 22500 }, { "epoch": 0.21, "grad_norm": 392361.15625, "learning_rate": 4.726161685158861e-05, "loss": 0.8802, "step": 23000 }, { "epoch": 0.22, "grad_norm": 844823.375, "learning_rate": 4.698469547879705e-05, "loss": 0.8927, "step": 23500 }, { "epoch": 0.22, "grad_norm": 252408.484375, "learning_rate": 4.6707774106005505e-05, "loss": 0.8942, "step": 24000 }, { "epoch": 0.23, "grad_norm": 591983.0625, "learning_rate": 4.643085273321395e-05, "loss": 0.8859, "step": 24500 }, { "epoch": 0.23, "grad_norm": 284245.46875, "learning_rate": 4.6153931360422396e-05, "loss": 0.8953, "step": 25000 }, { "epoch": 0.24, "grad_norm": 288328.15625, "learning_rate": 4.587700998763085e-05, "loss": 0.8731, "step": 25500 }, { "epoch": 0.24, "grad_norm": 422104.84375, "learning_rate": 4.560008861483929e-05, "loss": 0.8695, "step": 26000 }, { "epoch": 0.24, "grad_norm": 450527.375, "learning_rate": 4.5323167242047745e-05, "loss": 0.8616, "step": 26500 }, { "epoch": 0.25, "grad_norm": 440466.5, "learning_rate": 4.504624586925619e-05, "loss": 0.8871, "step": 27000 }, { "epoch": 0.25, "grad_norm": 386183.34375, "learning_rate": 4.4769324496464636e-05, "loss": 0.8645, "step": 27500 }, { "epoch": 0.26, "grad_norm": 542583.0, "learning_rate": 4.449240312367309e-05, "loss": 0.8755, "step": 28000 }, { "epoch": 0.26, "grad_norm": 497213.90625, "learning_rate": 4.421548175088153e-05, "loss": 0.8663, "step": 28500 }, { "epoch": 0.27, "grad_norm": 1084880.375, "learning_rate": 4.3938560378089985e-05, "loss": 0.8371, "step": 29000 }, { "epoch": 0.27, "grad_norm": 365434.125, "learning_rate": 4.366163900529843e-05, "loss": 0.8791, "step": 29500 }, { "epoch": 0.28, "grad_norm": 310639.5, "learning_rate": 4.3384717632506876e-05, "loss": 0.8298, "step": 30000 }, { "epoch": 0.28, "grad_norm": 1322288.0, "learning_rate": 4.310779625971533e-05, "loss": 0.8658, "step": 30500 }, { "epoch": 0.29, "grad_norm": 488033.0625, "learning_rate": 4.283087488692377e-05, "loss": 0.8556, "step": 31000 }, { "epoch": 0.29, "grad_norm": 404639.3125, "learning_rate": 4.2553953514132225e-05, "loss": 0.8547, "step": 31500 }, { "epoch": 0.3, "grad_norm": 324229.59375, "learning_rate": 4.227703214134067e-05, "loss": 0.8853, "step": 32000 }, { "epoch": 0.3, "grad_norm": 391828.6875, "learning_rate": 4.2000110768549116e-05, "loss": 0.8725, "step": 32500 }, { "epoch": 0.3, "grad_norm": 481346.0, "learning_rate": 4.172318939575757e-05, "loss": 0.8484, "step": 33000 }, { "epoch": 0.31, "grad_norm": 500705.90625, "learning_rate": 4.144626802296601e-05, "loss": 0.8415, "step": 33500 }, { "epoch": 0.31, "grad_norm": 484916.0625, "learning_rate": 4.116934665017446e-05, "loss": 0.8464, "step": 34000 }, { "epoch": 0.32, "grad_norm": 305320.15625, "learning_rate": 4.089242527738291e-05, "loss": 0.8439, "step": 34500 }, { "epoch": 0.32, "grad_norm": 275214.6875, "learning_rate": 4.0615503904591356e-05, "loss": 0.8257, "step": 35000 }, { "epoch": 0.33, "grad_norm": 364166.0, "learning_rate": 4.033858253179981e-05, "loss": 0.8763, "step": 35500 }, { "epoch": 0.33, "grad_norm": 285496.46875, "learning_rate": 4.006166115900825e-05, "loss": 0.856, "step": 36000 }, { "epoch": 0.34, "grad_norm": 959005.0625, "learning_rate": 3.97847397862167e-05, "loss": 0.8306, "step": 36500 }, { "epoch": 0.34, "grad_norm": 357090.28125, "learning_rate": 3.950781841342515e-05, "loss": 0.8564, "step": 37000 }, { "epoch": 0.35, "grad_norm": 605763.8125, "learning_rate": 3.9230897040633596e-05, "loss": 0.8476, "step": 37500 }, { "epoch": 0.35, "grad_norm": 418555.625, "learning_rate": 3.895397566784205e-05, "loss": 0.857, "step": 38000 }, { "epoch": 0.36, "grad_norm": 726787.125, "learning_rate": 3.8677054295050493e-05, "loss": 0.8627, "step": 38500 }, { "epoch": 0.36, "grad_norm": 212226.21875, "learning_rate": 3.840013292225894e-05, "loss": 0.8172, "step": 39000 }, { "epoch": 0.36, "grad_norm": 219878.765625, "learning_rate": 3.812321154946739e-05, "loss": 0.8476, "step": 39500 }, { "epoch": 0.37, "grad_norm": 520633.34375, "learning_rate": 3.7846290176675836e-05, "loss": 0.8226, "step": 40000 }, { "epoch": 0.37, "grad_norm": 317160.71875, "learning_rate": 3.756936880388429e-05, "loss": 0.8235, "step": 40500 }, { "epoch": 0.38, "grad_norm": 409949.8125, "learning_rate": 3.7292447431092734e-05, "loss": 0.8112, "step": 41000 }, { "epoch": 0.38, "grad_norm": 287649.84375, "learning_rate": 3.701552605830118e-05, "loss": 0.8357, "step": 41500 }, { "epoch": 0.39, "grad_norm": 386562.09375, "learning_rate": 3.673860468550963e-05, "loss": 0.8425, "step": 42000 }, { "epoch": 0.39, "grad_norm": 454850.125, "learning_rate": 3.6461683312718076e-05, "loss": 0.8347, "step": 42500 }, { "epoch": 0.4, "grad_norm": 239857.5, "learning_rate": 3.618476193992653e-05, "loss": 0.8619, "step": 43000 }, { "epoch": 0.4, "grad_norm": 227077.171875, "learning_rate": 3.5907840567134974e-05, "loss": 0.847, "step": 43500 }, { "epoch": 0.41, "grad_norm": 465131.4375, "learning_rate": 3.563091919434342e-05, "loss": 0.8333, "step": 44000 }, { "epoch": 0.41, "grad_norm": 329121.21875, "learning_rate": 3.535399782155187e-05, "loss": 0.8466, "step": 44500 }, { "epoch": 0.42, "grad_norm": 362261.40625, "learning_rate": 3.5077076448760316e-05, "loss": 0.8076, "step": 45000 }, { "epoch": 0.42, "grad_norm": 729279.75, "learning_rate": 3.480015507596876e-05, "loss": 0.8015, "step": 45500 }, { "epoch": 0.42, "grad_norm": 311664.15625, "learning_rate": 3.4523233703177214e-05, "loss": 0.838, "step": 46000 }, { "epoch": 0.43, "grad_norm": 267956.34375, "learning_rate": 3.424631233038566e-05, "loss": 0.8229, "step": 46500 }, { "epoch": 0.43, "grad_norm": 393154.1875, "learning_rate": 3.396939095759411e-05, "loss": 0.7973, "step": 47000 }, { "epoch": 0.44, "grad_norm": 298007.96875, "learning_rate": 3.3692469584802556e-05, "loss": 0.8309, "step": 47500 }, { "epoch": 0.44, "grad_norm": 235954.09375, "learning_rate": 3.3415548212011e-05, "loss": 0.808, "step": 48000 }, { "epoch": 0.45, "grad_norm": 565394.3125, "learning_rate": 3.3138626839219454e-05, "loss": 0.8123, "step": 48500 }, { "epoch": 0.45, "grad_norm": 228396.78125, "learning_rate": 3.28617054664279e-05, "loss": 0.8456, "step": 49000 }, { "epoch": 0.46, "grad_norm": 314704.09375, "learning_rate": 3.258478409363635e-05, "loss": 0.8116, "step": 49500 }, { "epoch": 0.46, "grad_norm": 250648.609375, "learning_rate": 3.2307862720844797e-05, "loss": 0.8284, "step": 50000 }, { "epoch": 0.47, "grad_norm": 409734.96875, "learning_rate": 3.203094134805324e-05, "loss": 0.8015, "step": 50500 }, { "epoch": 0.47, "grad_norm": 171156.265625, "learning_rate": 3.1754019975261694e-05, "loss": 0.8322, "step": 51000 }, { "epoch": 0.48, "grad_norm": 300950.65625, "learning_rate": 3.147709860247014e-05, "loss": 0.8215, "step": 51500 }, { "epoch": 0.48, "grad_norm": 223661.734375, "learning_rate": 3.120017722967859e-05, "loss": 0.8004, "step": 52000 }, { "epoch": 0.48, "grad_norm": 303242.25, "learning_rate": 3.0923255856887037e-05, "loss": 0.8109, "step": 52500 }, { "epoch": 0.49, "grad_norm": 190445.625, "learning_rate": 3.064633448409548e-05, "loss": 0.8126, "step": 53000 }, { "epoch": 0.49, "grad_norm": 312735.21875, "learning_rate": 3.036941311130393e-05, "loss": 0.8186, "step": 53500 }, { "epoch": 0.5, "grad_norm": 492028.34375, "learning_rate": 3.009249173851238e-05, "loss": 0.7993, "step": 54000 }, { "epoch": 0.5, "grad_norm": 159610.984375, "learning_rate": 2.9815570365720828e-05, "loss": 0.8209, "step": 54500 }, { "epoch": 0.51, "grad_norm": 361054.71875, "learning_rate": 2.9538648992929277e-05, "loss": 0.8034, "step": 55000 }, { "epoch": 0.51, "grad_norm": 333324.875, "learning_rate": 2.9261727620137725e-05, "loss": 0.7747, "step": 55500 }, { "epoch": 0.52, "grad_norm": 220997.15625, "learning_rate": 2.898480624734617e-05, "loss": 0.7898, "step": 56000 }, { "epoch": 0.52, "grad_norm": 427477.28125, "learning_rate": 2.870788487455462e-05, "loss": 0.8487, "step": 56500 }, { "epoch": 0.53, "grad_norm": 494515.15625, "learning_rate": 2.8430963501763068e-05, "loss": 0.7867, "step": 57000 }, { "epoch": 0.53, "grad_norm": 131962.8125, "learning_rate": 2.8154042128971517e-05, "loss": 0.8257, "step": 57500 }, { "epoch": 0.54, "grad_norm": 360923.75, "learning_rate": 2.7877120756179962e-05, "loss": 0.824, "step": 58000 }, { "epoch": 0.54, "grad_norm": 537102.3125, "learning_rate": 2.760019938338841e-05, "loss": 0.7947, "step": 58500 }, { "epoch": 0.54, "grad_norm": 185466.890625, "learning_rate": 2.732327801059686e-05, "loss": 0.7965, "step": 59000 }, { "epoch": 0.55, "grad_norm": 449036.0, "learning_rate": 2.7046356637805308e-05, "loss": 0.8359, "step": 59500 }, { "epoch": 0.55, "grad_norm": 302378.875, "learning_rate": 2.6769435265013757e-05, "loss": 0.791, "step": 60000 }, { "epoch": 0.56, "grad_norm": 170428.640625, "learning_rate": 2.6492513892222202e-05, "loss": 0.804, "step": 60500 }, { "epoch": 0.56, "grad_norm": 486227.0, "learning_rate": 2.621559251943065e-05, "loss": 0.7807, "step": 61000 }, { "epoch": 0.57, "grad_norm": 657373.0, "learning_rate": 2.59386711466391e-05, "loss": 0.8004, "step": 61500 }, { "epoch": 0.57, "grad_norm": 386137.0625, "learning_rate": 2.5661749773847548e-05, "loss": 0.7669, "step": 62000 }, { "epoch": 0.58, "grad_norm": 451425.8125, "learning_rate": 2.5384828401055994e-05, "loss": 0.7948, "step": 62500 }, { "epoch": 0.58, "grad_norm": 188410.265625, "learning_rate": 2.5107907028264442e-05, "loss": 0.7988, "step": 63000 }, { "epoch": 0.59, "grad_norm": 543696.4375, "learning_rate": 2.483098565547289e-05, "loss": 0.7927, "step": 63500 }, { "epoch": 0.59, "grad_norm": 317845.90625, "learning_rate": 2.455406428268134e-05, "loss": 0.8021, "step": 64000 }, { "epoch": 0.6, "grad_norm": 567647.5, "learning_rate": 2.427714290988979e-05, "loss": 0.7853, "step": 64500 }, { "epoch": 0.6, "grad_norm": 413894.4375, "learning_rate": 2.4000221537098234e-05, "loss": 0.7973, "step": 65000 }, { "epoch": 0.6, "grad_norm": 519043.9375, "learning_rate": 2.3723300164306682e-05, "loss": 0.7516, "step": 65500 }, { "epoch": 0.61, "grad_norm": 129269.5234375, "learning_rate": 2.344637879151513e-05, "loss": 0.8054, "step": 66000 }, { "epoch": 0.61, "grad_norm": 272380.0625, "learning_rate": 2.316945741872358e-05, "loss": 0.8112, "step": 66500 }, { "epoch": 0.62, "grad_norm": 299374.875, "learning_rate": 2.2892536045932025e-05, "loss": 0.7687, "step": 67000 }, { "epoch": 0.62, "grad_norm": 597694.5, "learning_rate": 2.2615614673140474e-05, "loss": 0.8072, "step": 67500 }, { "epoch": 0.63, "grad_norm": 273690.375, "learning_rate": 2.2338693300348922e-05, "loss": 0.7895, "step": 68000 }, { "epoch": 0.63, "grad_norm": 237268.21875, "learning_rate": 2.206177192755737e-05, "loss": 0.7802, "step": 68500 }, { "epoch": 0.64, "grad_norm": 347646.09375, "learning_rate": 2.178485055476582e-05, "loss": 0.7991, "step": 69000 }, { "epoch": 0.64, "grad_norm": 266270.28125, "learning_rate": 2.1507929181974265e-05, "loss": 0.7979, "step": 69500 }, { "epoch": 0.65, "grad_norm": 371650.84375, "learning_rate": 2.1231007809182714e-05, "loss": 0.7616, "step": 70000 }, { "epoch": 0.65, "grad_norm": 389250.75, "learning_rate": 2.0954086436391163e-05, "loss": 0.8056, "step": 70500 }, { "epoch": 0.66, "grad_norm": 472425.71875, "learning_rate": 2.067716506359961e-05, "loss": 0.7632, "step": 71000 }, { "epoch": 0.66, "grad_norm": 341248.3125, "learning_rate": 2.0400243690808057e-05, "loss": 0.7648, "step": 71500 }, { "epoch": 0.66, "grad_norm": 470803.96875, "learning_rate": 2.0123322318016505e-05, "loss": 0.7699, "step": 72000 }, { "epoch": 0.67, "grad_norm": 363643.71875, "learning_rate": 1.9846400945224954e-05, "loss": 0.7929, "step": 72500 }, { "epoch": 0.67, "grad_norm": 380603.65625, "learning_rate": 1.9569479572433403e-05, "loss": 0.7614, "step": 73000 }, { "epoch": 0.68, "grad_norm": 282191.78125, "learning_rate": 1.929255819964185e-05, "loss": 0.7508, "step": 73500 }, { "epoch": 0.68, "grad_norm": 270537.5625, "learning_rate": 1.9015636826850297e-05, "loss": 0.7724, "step": 74000 }, { "epoch": 0.69, "grad_norm": 786935.9375, "learning_rate": 1.8738715454058745e-05, "loss": 0.7903, "step": 74500 }, { "epoch": 0.69, "grad_norm": 276354.875, "learning_rate": 1.8461794081267194e-05, "loss": 0.8034, "step": 75000 }, { "epoch": 0.7, "grad_norm": 488051.25, "learning_rate": 1.8184872708475643e-05, "loss": 0.7988, "step": 75500 }, { "epoch": 0.7, "grad_norm": 268592.53125, "learning_rate": 1.7907951335684088e-05, "loss": 0.771, "step": 76000 }, { "epoch": 0.71, "grad_norm": 286408.90625, "learning_rate": 1.7631029962892537e-05, "loss": 0.7732, "step": 76500 }, { "epoch": 0.71, "grad_norm": 287447.375, "learning_rate": 1.7354108590100985e-05, "loss": 0.7788, "step": 77000 }, { "epoch": 0.72, "grad_norm": 427488.15625, "learning_rate": 1.7077187217309434e-05, "loss": 0.7763, "step": 77500 }, { "epoch": 0.72, "grad_norm": 327204.96875, "learning_rate": 1.6800265844517883e-05, "loss": 0.7814, "step": 78000 }, { "epoch": 0.72, "grad_norm": 446818.0, "learning_rate": 1.6523344471726328e-05, "loss": 0.773, "step": 78500 }, { "epoch": 0.73, "grad_norm": 367178.46875, "learning_rate": 1.6246423098934777e-05, "loss": 0.7784, "step": 79000 }, { "epoch": 0.73, "grad_norm": 564747.75, "learning_rate": 1.5969501726143226e-05, "loss": 0.7751, "step": 79500 }, { "epoch": 0.74, "grad_norm": 373075.34375, "learning_rate": 1.5692580353351674e-05, "loss": 0.7701, "step": 80000 }, { "epoch": 0.74, "grad_norm": 220253.890625, "learning_rate": 1.541565898056012e-05, "loss": 0.7717, "step": 80500 }, { "epoch": 0.75, "grad_norm": 442242.46875, "learning_rate": 1.513873760776857e-05, "loss": 0.7699, "step": 81000 }, { "epoch": 0.75, "grad_norm": 281603.1875, "learning_rate": 1.4861816234977015e-05, "loss": 0.7626, "step": 81500 }, { "epoch": 0.76, "grad_norm": 420246.21875, "learning_rate": 1.4584894862185464e-05, "loss": 0.7719, "step": 82000 }, { "epoch": 0.76, "grad_norm": 347264.375, "learning_rate": 1.4307973489393911e-05, "loss": 0.764, "step": 82500 }, { "epoch": 0.77, "grad_norm": 326494.21875, "learning_rate": 1.403105211660236e-05, "loss": 0.7617, "step": 83000 }, { "epoch": 0.77, "grad_norm": 193308.671875, "learning_rate": 1.3754130743810807e-05, "loss": 0.7599, "step": 83500 }, { "epoch": 0.78, "grad_norm": 283088.8125, "learning_rate": 1.3477209371019255e-05, "loss": 0.7628, "step": 84000 }, { "epoch": 0.78, "grad_norm": 539284.0, "learning_rate": 1.3200287998227702e-05, "loss": 0.7358, "step": 84500 }, { "epoch": 0.78, "grad_norm": 473964.84375, "learning_rate": 1.2923366625436151e-05, "loss": 0.755, "step": 85000 }, { "epoch": 0.79, "grad_norm": 327867.3125, "learning_rate": 1.26464452526446e-05, "loss": 0.7718, "step": 85500 }, { "epoch": 0.79, "grad_norm": 263024.28125, "learning_rate": 1.2369523879853047e-05, "loss": 0.7722, "step": 86000 }, { "epoch": 0.8, "grad_norm": 266528.5625, "learning_rate": 1.2092602507061495e-05, "loss": 0.7635, "step": 86500 }, { "epoch": 0.8, "grad_norm": 126082.46875, "learning_rate": 1.1815681134269942e-05, "loss": 0.7508, "step": 87000 }, { "epoch": 0.81, "grad_norm": 300010.3125, "learning_rate": 1.1538759761478391e-05, "loss": 0.7755, "step": 87500 }, { "epoch": 0.81, "grad_norm": 471967.5625, "learning_rate": 1.1261838388686838e-05, "loss": 0.7416, "step": 88000 }, { "epoch": 0.82, "grad_norm": 261213.0625, "learning_rate": 1.0984917015895287e-05, "loss": 0.7705, "step": 88500 }, { "epoch": 0.82, "grad_norm": 192903.46875, "learning_rate": 1.0707995643103734e-05, "loss": 0.7392, "step": 89000 }, { "epoch": 0.83, "grad_norm": 682632.8125, "learning_rate": 1.0431074270312182e-05, "loss": 0.7577, "step": 89500 }, { "epoch": 0.83, "grad_norm": 218102.4375, "learning_rate": 1.0154152897520631e-05, "loss": 0.7586, "step": 90000 }, { "epoch": 0.84, "grad_norm": 212267.5625, "learning_rate": 9.877231524729078e-06, "loss": 0.7472, "step": 90500 }, { "epoch": 0.84, "grad_norm": 320534.59375, "learning_rate": 9.600310151937527e-06, "loss": 0.7602, "step": 91000 }, { "epoch": 0.84, "grad_norm": 295994.96875, "learning_rate": 9.323388779145974e-06, "loss": 0.7368, "step": 91500 }, { "epoch": 0.85, "grad_norm": 221153.6875, "learning_rate": 9.046467406354423e-06, "loss": 0.7492, "step": 92000 }, { "epoch": 0.85, "grad_norm": 250322.6875, "learning_rate": 8.76954603356287e-06, "loss": 0.7582, "step": 92500 }, { "epoch": 0.86, "grad_norm": 364772.46875, "learning_rate": 8.492624660771318e-06, "loss": 0.747, "step": 93000 }, { "epoch": 0.86, "grad_norm": 226077.453125, "learning_rate": 8.215703287979767e-06, "loss": 0.7515, "step": 93500 }, { "epoch": 0.87, "grad_norm": 260477.265625, "learning_rate": 7.938781915188214e-06, "loss": 0.7402, "step": 94000 }, { "epoch": 0.87, "grad_norm": 380977.46875, "learning_rate": 7.661860542396663e-06, "loss": 0.7512, "step": 94500 }, { "epoch": 0.88, "grad_norm": 1052145.875, "learning_rate": 7.38493916960511e-06, "loss": 0.7589, "step": 95000 }, { "epoch": 0.88, "grad_norm": 213601.171875, "learning_rate": 7.1080177968135575e-06, "loss": 0.7581, "step": 95500 }, { "epoch": 0.89, "grad_norm": 467877.71875, "learning_rate": 6.831096424022006e-06, "loss": 0.7464, "step": 96000 }, { "epoch": 0.89, "grad_norm": 566204.3125, "learning_rate": 6.554175051230454e-06, "loss": 0.7557, "step": 96500 }, { "epoch": 0.9, "grad_norm": 330177.5625, "learning_rate": 6.277253678438902e-06, "loss": 0.7304, "step": 97000 }, { "epoch": 0.9, "grad_norm": 319206.65625, "learning_rate": 6.00033230564735e-06, "loss": 0.736, "step": 97500 }, { "epoch": 0.9, "grad_norm": 410904.90625, "learning_rate": 5.723410932855798e-06, "loss": 0.752, "step": 98000 }, { "epoch": 0.91, "grad_norm": 375357.15625, "learning_rate": 5.4464895600642454e-06, "loss": 0.7358, "step": 98500 }, { "epoch": 0.91, "grad_norm": 307483.375, "learning_rate": 5.169568187272693e-06, "loss": 0.7749, "step": 99000 }, { "epoch": 0.92, "grad_norm": 434196.59375, "learning_rate": 4.892646814481141e-06, "loss": 0.7324, "step": 99500 }, { "epoch": 0.92, "grad_norm": 554119.4375, "learning_rate": 4.61572544168959e-06, "loss": 0.7096, "step": 100000 }, { "epoch": 0.93, "grad_norm": 224006.78125, "learning_rate": 4.338804068898038e-06, "loss": 0.7205, "step": 100500 }, { "epoch": 0.93, "grad_norm": 215164.78125, "learning_rate": 4.0618826961064855e-06, "loss": 0.732, "step": 101000 }, { "epoch": 0.94, "grad_norm": 391820.53125, "learning_rate": 3.7849613233149334e-06, "loss": 0.7228, "step": 101500 }, { "epoch": 0.94, "grad_norm": 306694.15625, "learning_rate": 3.5080399505233816e-06, "loss": 0.7507, "step": 102000 }, { "epoch": 0.95, "grad_norm": 467995.03125, "learning_rate": 3.2311185777318295e-06, "loss": 0.7433, "step": 102500 }, { "epoch": 0.95, "grad_norm": 284427.625, "learning_rate": 2.9541972049402773e-06, "loss": 0.7047, "step": 103000 }, { "epoch": 0.96, "grad_norm": 193486.921875, "learning_rate": 2.6772758321487256e-06, "loss": 0.7646, "step": 103500 }, { "epoch": 0.96, "grad_norm": 262856.15625, "learning_rate": 2.4003544593571735e-06, "loss": 0.7306, "step": 104000 }, { "epoch": 0.96, "grad_norm": 267263.5, "learning_rate": 2.1234330865656213e-06, "loss": 0.7631, "step": 104500 }, { "epoch": 0.97, "grad_norm": 229738.515625, "learning_rate": 1.8465117137740692e-06, "loss": 0.731, "step": 105000 }, { "epoch": 0.97, "grad_norm": 528997.6875, "learning_rate": 1.569590340982517e-06, "loss": 0.7393, "step": 105500 }, { "epoch": 0.98, "grad_norm": 346399.28125, "learning_rate": 1.292668968190965e-06, "loss": 0.7304, "step": 106000 }, { "epoch": 0.98, "grad_norm": 315753.28125, "learning_rate": 1.015747595399413e-06, "loss": 0.7239, "step": 106500 }, { "epoch": 0.99, "grad_norm": 245733.109375, "learning_rate": 7.388262226078609e-07, "loss": 0.7067, "step": 107000 }, { "epoch": 0.99, "grad_norm": 423298.15625, "learning_rate": 4.619048498163089e-07, "loss": 0.734, "step": 107500 }, { "epoch": 1.0, "grad_norm": 361011.8125, "learning_rate": 1.8498347702475678e-07, "loss": 0.7356, "step": 108000 }, { "epoch": 1.0, "eval_accuracy": 0.68, "eval_loss": 0.750106155872345, "eval_runtime": 570.5256, "eval_samples_per_second": 17.528, "eval_steps_per_second": 2.922, "step": 108334 } ], "logging_steps": 500, "max_steps": 108334, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "total_flos": 2.328341808688051e+18, "train_batch_size": 6, "trial_name": null, "trial_params": null }