| { | |
| "best_metric": 0.750106155872345, | |
| "best_model_checkpoint": "sean_test_out_large/checkpoint-108334", | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 108334, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 1387778.25, | |
| "learning_rate": 5.972307862720845e-05, | |
| "loss": 1.1897, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 664302.625, | |
| "learning_rate": 5.94461572544169e-05, | |
| "loss": 1.0759, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 428486.84375, | |
| "learning_rate": 5.9169235881625344e-05, | |
| "loss": 1.0728, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 299683.625, | |
| "learning_rate": 5.8892314508833796e-05, | |
| "loss": 1.0409, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 503517.40625, | |
| "learning_rate": 5.861539313604224e-05, | |
| "loss": 1.0335, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 657503.8125, | |
| "learning_rate": 5.833847176325069e-05, | |
| "loss": 0.9533, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 533787.3125, | |
| "learning_rate": 5.806155039045914e-05, | |
| "loss": 0.9992, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 913784.6875, | |
| "learning_rate": 5.7784629017667584e-05, | |
| "loss": 0.9657, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 310152.21875, | |
| "learning_rate": 5.7507707644876036e-05, | |
| "loss": 0.9652, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 309690.34375, | |
| "learning_rate": 5.723078627208448e-05, | |
| "loss": 0.969, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 211456.5, | |
| "learning_rate": 5.695386489929293e-05, | |
| "loss": 0.9515, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 391416.65625, | |
| "learning_rate": 5.667694352650138e-05, | |
| "loss": 0.9221, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 221054.703125, | |
| "learning_rate": 5.6400022153709824e-05, | |
| "loss": 0.9679, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 472076.15625, | |
| "learning_rate": 5.612310078091827e-05, | |
| "loss": 0.9519, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 988913.375, | |
| "learning_rate": 5.584617940812672e-05, | |
| "loss": 0.9447, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 661127.75, | |
| "learning_rate": 5.556925803533517e-05, | |
| "loss": 0.9009, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 212504.640625, | |
| "learning_rate": 5.529233666254362e-05, | |
| "loss": 0.9456, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 473405.96875, | |
| "learning_rate": 5.5015415289752064e-05, | |
| "loss": 0.9385, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 266084.78125, | |
| "learning_rate": 5.473849391696051e-05, | |
| "loss": 1.0131, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 491548.53125, | |
| "learning_rate": 5.446157254416896e-05, | |
| "loss": 0.9254, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 267938.4375, | |
| "learning_rate": 5.418465117137741e-05, | |
| "loss": 0.9101, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 258540.84375, | |
| "learning_rate": 5.390772979858586e-05, | |
| "loss": 0.9234, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 200250.609375, | |
| "learning_rate": 5.3630808425794305e-05, | |
| "loss": 0.9157, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 480926.25, | |
| "learning_rate": 5.335388705300275e-05, | |
| "loss": 0.9532, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 593979.4375, | |
| "learning_rate": 5.30769656802112e-05, | |
| "loss": 0.9184, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 686144.1875, | |
| "learning_rate": 5.280004430741965e-05, | |
| "loss": 0.9215, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 582782.0, | |
| "learning_rate": 5.25231229346281e-05, | |
| "loss": 0.9151, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 554729.875, | |
| "learning_rate": 5.2246201561836545e-05, | |
| "loss": 0.9042, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 423190.0, | |
| "learning_rate": 5.196928018904499e-05, | |
| "loss": 0.9114, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 724497.5, | |
| "learning_rate": 5.169235881625344e-05, | |
| "loss": 0.9123, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 333863.78125, | |
| "learning_rate": 5.141543744346189e-05, | |
| "loss": 0.9237, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 282522.5, | |
| "learning_rate": 5.113851607067033e-05, | |
| "loss": 0.8927, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 326572.34375, | |
| "learning_rate": 5.0861594697878785e-05, | |
| "loss": 0.9256, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 454635.34375, | |
| "learning_rate": 5.058467332508723e-05, | |
| "loss": 0.8991, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 306963.25, | |
| "learning_rate": 5.030775195229568e-05, | |
| "loss": 0.8937, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 580224.0, | |
| "learning_rate": 5.003083057950413e-05, | |
| "loss": 0.9108, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 549220.75, | |
| "learning_rate": 4.975390920671257e-05, | |
| "loss": 0.9307, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 188370.0, | |
| "learning_rate": 4.9476987833921025e-05, | |
| "loss": 0.9009, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 535151.8125, | |
| "learning_rate": 4.920006646112947e-05, | |
| "loss": 0.8773, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 675718.1875, | |
| "learning_rate": 4.892314508833792e-05, | |
| "loss": 0.8984, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 247562.0, | |
| "learning_rate": 4.864622371554637e-05, | |
| "loss": 0.8896, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 412286.65625, | |
| "learning_rate": 4.836930234275481e-05, | |
| "loss": 0.8914, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 362339.53125, | |
| "learning_rate": 4.8092380969963265e-05, | |
| "loss": 0.8906, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 307259.8125, | |
| "learning_rate": 4.781545959717171e-05, | |
| "loss": 0.8615, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 264488.625, | |
| "learning_rate": 4.753853822438016e-05, | |
| "loss": 0.896, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 392361.15625, | |
| "learning_rate": 4.726161685158861e-05, | |
| "loss": 0.8802, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 844823.375, | |
| "learning_rate": 4.698469547879705e-05, | |
| "loss": 0.8927, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 252408.484375, | |
| "learning_rate": 4.6707774106005505e-05, | |
| "loss": 0.8942, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 591983.0625, | |
| "learning_rate": 4.643085273321395e-05, | |
| "loss": 0.8859, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 284245.46875, | |
| "learning_rate": 4.6153931360422396e-05, | |
| "loss": 0.8953, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 288328.15625, | |
| "learning_rate": 4.587700998763085e-05, | |
| "loss": 0.8731, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 422104.84375, | |
| "learning_rate": 4.560008861483929e-05, | |
| "loss": 0.8695, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 450527.375, | |
| "learning_rate": 4.5323167242047745e-05, | |
| "loss": 0.8616, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 440466.5, | |
| "learning_rate": 4.504624586925619e-05, | |
| "loss": 0.8871, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 386183.34375, | |
| "learning_rate": 4.4769324496464636e-05, | |
| "loss": 0.8645, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 542583.0, | |
| "learning_rate": 4.449240312367309e-05, | |
| "loss": 0.8755, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 497213.90625, | |
| "learning_rate": 4.421548175088153e-05, | |
| "loss": 0.8663, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 1084880.375, | |
| "learning_rate": 4.3938560378089985e-05, | |
| "loss": 0.8371, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 365434.125, | |
| "learning_rate": 4.366163900529843e-05, | |
| "loss": 0.8791, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 310639.5, | |
| "learning_rate": 4.3384717632506876e-05, | |
| "loss": 0.8298, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 1322288.0, | |
| "learning_rate": 4.310779625971533e-05, | |
| "loss": 0.8658, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 488033.0625, | |
| "learning_rate": 4.283087488692377e-05, | |
| "loss": 0.8556, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 404639.3125, | |
| "learning_rate": 4.2553953514132225e-05, | |
| "loss": 0.8547, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 324229.59375, | |
| "learning_rate": 4.227703214134067e-05, | |
| "loss": 0.8853, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 391828.6875, | |
| "learning_rate": 4.2000110768549116e-05, | |
| "loss": 0.8725, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 481346.0, | |
| "learning_rate": 4.172318939575757e-05, | |
| "loss": 0.8484, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 500705.90625, | |
| "learning_rate": 4.144626802296601e-05, | |
| "loss": 0.8415, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 484916.0625, | |
| "learning_rate": 4.116934665017446e-05, | |
| "loss": 0.8464, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 305320.15625, | |
| "learning_rate": 4.089242527738291e-05, | |
| "loss": 0.8439, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 275214.6875, | |
| "learning_rate": 4.0615503904591356e-05, | |
| "loss": 0.8257, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 364166.0, | |
| "learning_rate": 4.033858253179981e-05, | |
| "loss": 0.8763, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 285496.46875, | |
| "learning_rate": 4.006166115900825e-05, | |
| "loss": 0.856, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 959005.0625, | |
| "learning_rate": 3.97847397862167e-05, | |
| "loss": 0.8306, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 357090.28125, | |
| "learning_rate": 3.950781841342515e-05, | |
| "loss": 0.8564, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 605763.8125, | |
| "learning_rate": 3.9230897040633596e-05, | |
| "loss": 0.8476, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 418555.625, | |
| "learning_rate": 3.895397566784205e-05, | |
| "loss": 0.857, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 726787.125, | |
| "learning_rate": 3.8677054295050493e-05, | |
| "loss": 0.8627, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 212226.21875, | |
| "learning_rate": 3.840013292225894e-05, | |
| "loss": 0.8172, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 219878.765625, | |
| "learning_rate": 3.812321154946739e-05, | |
| "loss": 0.8476, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 520633.34375, | |
| "learning_rate": 3.7846290176675836e-05, | |
| "loss": 0.8226, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 317160.71875, | |
| "learning_rate": 3.756936880388429e-05, | |
| "loss": 0.8235, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 409949.8125, | |
| "learning_rate": 3.7292447431092734e-05, | |
| "loss": 0.8112, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 287649.84375, | |
| "learning_rate": 3.701552605830118e-05, | |
| "loss": 0.8357, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 386562.09375, | |
| "learning_rate": 3.673860468550963e-05, | |
| "loss": 0.8425, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 454850.125, | |
| "learning_rate": 3.6461683312718076e-05, | |
| "loss": 0.8347, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 239857.5, | |
| "learning_rate": 3.618476193992653e-05, | |
| "loss": 0.8619, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 227077.171875, | |
| "learning_rate": 3.5907840567134974e-05, | |
| "loss": 0.847, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 465131.4375, | |
| "learning_rate": 3.563091919434342e-05, | |
| "loss": 0.8333, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 329121.21875, | |
| "learning_rate": 3.535399782155187e-05, | |
| "loss": 0.8466, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 362261.40625, | |
| "learning_rate": 3.5077076448760316e-05, | |
| "loss": 0.8076, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 729279.75, | |
| "learning_rate": 3.480015507596876e-05, | |
| "loss": 0.8015, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 311664.15625, | |
| "learning_rate": 3.4523233703177214e-05, | |
| "loss": 0.838, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 267956.34375, | |
| "learning_rate": 3.424631233038566e-05, | |
| "loss": 0.8229, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 393154.1875, | |
| "learning_rate": 3.396939095759411e-05, | |
| "loss": 0.7973, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 298007.96875, | |
| "learning_rate": 3.3692469584802556e-05, | |
| "loss": 0.8309, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 235954.09375, | |
| "learning_rate": 3.3415548212011e-05, | |
| "loss": 0.808, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 565394.3125, | |
| "learning_rate": 3.3138626839219454e-05, | |
| "loss": 0.8123, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 228396.78125, | |
| "learning_rate": 3.28617054664279e-05, | |
| "loss": 0.8456, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 314704.09375, | |
| "learning_rate": 3.258478409363635e-05, | |
| "loss": 0.8116, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 250648.609375, | |
| "learning_rate": 3.2307862720844797e-05, | |
| "loss": 0.8284, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 409734.96875, | |
| "learning_rate": 3.203094134805324e-05, | |
| "loss": 0.8015, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 171156.265625, | |
| "learning_rate": 3.1754019975261694e-05, | |
| "loss": 0.8322, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 300950.65625, | |
| "learning_rate": 3.147709860247014e-05, | |
| "loss": 0.8215, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 223661.734375, | |
| "learning_rate": 3.120017722967859e-05, | |
| "loss": 0.8004, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 303242.25, | |
| "learning_rate": 3.0923255856887037e-05, | |
| "loss": 0.8109, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 190445.625, | |
| "learning_rate": 3.064633448409548e-05, | |
| "loss": 0.8126, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 312735.21875, | |
| "learning_rate": 3.036941311130393e-05, | |
| "loss": 0.8186, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 492028.34375, | |
| "learning_rate": 3.009249173851238e-05, | |
| "loss": 0.7993, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 159610.984375, | |
| "learning_rate": 2.9815570365720828e-05, | |
| "loss": 0.8209, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 361054.71875, | |
| "learning_rate": 2.9538648992929277e-05, | |
| "loss": 0.8034, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 333324.875, | |
| "learning_rate": 2.9261727620137725e-05, | |
| "loss": 0.7747, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 220997.15625, | |
| "learning_rate": 2.898480624734617e-05, | |
| "loss": 0.7898, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 427477.28125, | |
| "learning_rate": 2.870788487455462e-05, | |
| "loss": 0.8487, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 494515.15625, | |
| "learning_rate": 2.8430963501763068e-05, | |
| "loss": 0.7867, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 131962.8125, | |
| "learning_rate": 2.8154042128971517e-05, | |
| "loss": 0.8257, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 360923.75, | |
| "learning_rate": 2.7877120756179962e-05, | |
| "loss": 0.824, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 537102.3125, | |
| "learning_rate": 2.760019938338841e-05, | |
| "loss": 0.7947, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 185466.890625, | |
| "learning_rate": 2.732327801059686e-05, | |
| "loss": 0.7965, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 449036.0, | |
| "learning_rate": 2.7046356637805308e-05, | |
| "loss": 0.8359, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 302378.875, | |
| "learning_rate": 2.6769435265013757e-05, | |
| "loss": 0.791, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 170428.640625, | |
| "learning_rate": 2.6492513892222202e-05, | |
| "loss": 0.804, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 486227.0, | |
| "learning_rate": 2.621559251943065e-05, | |
| "loss": 0.7807, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 657373.0, | |
| "learning_rate": 2.59386711466391e-05, | |
| "loss": 0.8004, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 386137.0625, | |
| "learning_rate": 2.5661749773847548e-05, | |
| "loss": 0.7669, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 451425.8125, | |
| "learning_rate": 2.5384828401055994e-05, | |
| "loss": 0.7948, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 188410.265625, | |
| "learning_rate": 2.5107907028264442e-05, | |
| "loss": 0.7988, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 543696.4375, | |
| "learning_rate": 2.483098565547289e-05, | |
| "loss": 0.7927, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 317845.90625, | |
| "learning_rate": 2.455406428268134e-05, | |
| "loss": 0.8021, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 567647.5, | |
| "learning_rate": 2.427714290988979e-05, | |
| "loss": 0.7853, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 413894.4375, | |
| "learning_rate": 2.4000221537098234e-05, | |
| "loss": 0.7973, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 519043.9375, | |
| "learning_rate": 2.3723300164306682e-05, | |
| "loss": 0.7516, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 129269.5234375, | |
| "learning_rate": 2.344637879151513e-05, | |
| "loss": 0.8054, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 272380.0625, | |
| "learning_rate": 2.316945741872358e-05, | |
| "loss": 0.8112, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 299374.875, | |
| "learning_rate": 2.2892536045932025e-05, | |
| "loss": 0.7687, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 597694.5, | |
| "learning_rate": 2.2615614673140474e-05, | |
| "loss": 0.8072, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 273690.375, | |
| "learning_rate": 2.2338693300348922e-05, | |
| "loss": 0.7895, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 237268.21875, | |
| "learning_rate": 2.206177192755737e-05, | |
| "loss": 0.7802, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 347646.09375, | |
| "learning_rate": 2.178485055476582e-05, | |
| "loss": 0.7991, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 266270.28125, | |
| "learning_rate": 2.1507929181974265e-05, | |
| "loss": 0.7979, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 371650.84375, | |
| "learning_rate": 2.1231007809182714e-05, | |
| "loss": 0.7616, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 389250.75, | |
| "learning_rate": 2.0954086436391163e-05, | |
| "loss": 0.8056, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 472425.71875, | |
| "learning_rate": 2.067716506359961e-05, | |
| "loss": 0.7632, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 341248.3125, | |
| "learning_rate": 2.0400243690808057e-05, | |
| "loss": 0.7648, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 470803.96875, | |
| "learning_rate": 2.0123322318016505e-05, | |
| "loss": 0.7699, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 363643.71875, | |
| "learning_rate": 1.9846400945224954e-05, | |
| "loss": 0.7929, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 380603.65625, | |
| "learning_rate": 1.9569479572433403e-05, | |
| "loss": 0.7614, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 282191.78125, | |
| "learning_rate": 1.929255819964185e-05, | |
| "loss": 0.7508, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 270537.5625, | |
| "learning_rate": 1.9015636826850297e-05, | |
| "loss": 0.7724, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 786935.9375, | |
| "learning_rate": 1.8738715454058745e-05, | |
| "loss": 0.7903, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 276354.875, | |
| "learning_rate": 1.8461794081267194e-05, | |
| "loss": 0.8034, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 488051.25, | |
| "learning_rate": 1.8184872708475643e-05, | |
| "loss": 0.7988, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 268592.53125, | |
| "learning_rate": 1.7907951335684088e-05, | |
| "loss": 0.771, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 286408.90625, | |
| "learning_rate": 1.7631029962892537e-05, | |
| "loss": 0.7732, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 287447.375, | |
| "learning_rate": 1.7354108590100985e-05, | |
| "loss": 0.7788, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 427488.15625, | |
| "learning_rate": 1.7077187217309434e-05, | |
| "loss": 0.7763, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 327204.96875, | |
| "learning_rate": 1.6800265844517883e-05, | |
| "loss": 0.7814, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 446818.0, | |
| "learning_rate": 1.6523344471726328e-05, | |
| "loss": 0.773, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 367178.46875, | |
| "learning_rate": 1.6246423098934777e-05, | |
| "loss": 0.7784, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 564747.75, | |
| "learning_rate": 1.5969501726143226e-05, | |
| "loss": 0.7751, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 373075.34375, | |
| "learning_rate": 1.5692580353351674e-05, | |
| "loss": 0.7701, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 220253.890625, | |
| "learning_rate": 1.541565898056012e-05, | |
| "loss": 0.7717, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 442242.46875, | |
| "learning_rate": 1.513873760776857e-05, | |
| "loss": 0.7699, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 281603.1875, | |
| "learning_rate": 1.4861816234977015e-05, | |
| "loss": 0.7626, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 420246.21875, | |
| "learning_rate": 1.4584894862185464e-05, | |
| "loss": 0.7719, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 347264.375, | |
| "learning_rate": 1.4307973489393911e-05, | |
| "loss": 0.764, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 326494.21875, | |
| "learning_rate": 1.403105211660236e-05, | |
| "loss": 0.7617, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 193308.671875, | |
| "learning_rate": 1.3754130743810807e-05, | |
| "loss": 0.7599, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 283088.8125, | |
| "learning_rate": 1.3477209371019255e-05, | |
| "loss": 0.7628, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 539284.0, | |
| "learning_rate": 1.3200287998227702e-05, | |
| "loss": 0.7358, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 473964.84375, | |
| "learning_rate": 1.2923366625436151e-05, | |
| "loss": 0.755, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 327867.3125, | |
| "learning_rate": 1.26464452526446e-05, | |
| "loss": 0.7718, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 263024.28125, | |
| "learning_rate": 1.2369523879853047e-05, | |
| "loss": 0.7722, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 266528.5625, | |
| "learning_rate": 1.2092602507061495e-05, | |
| "loss": 0.7635, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 126082.46875, | |
| "learning_rate": 1.1815681134269942e-05, | |
| "loss": 0.7508, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 300010.3125, | |
| "learning_rate": 1.1538759761478391e-05, | |
| "loss": 0.7755, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 471967.5625, | |
| "learning_rate": 1.1261838388686838e-05, | |
| "loss": 0.7416, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 261213.0625, | |
| "learning_rate": 1.0984917015895287e-05, | |
| "loss": 0.7705, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 192903.46875, | |
| "learning_rate": 1.0707995643103734e-05, | |
| "loss": 0.7392, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 682632.8125, | |
| "learning_rate": 1.0431074270312182e-05, | |
| "loss": 0.7577, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 218102.4375, | |
| "learning_rate": 1.0154152897520631e-05, | |
| "loss": 0.7586, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 212267.5625, | |
| "learning_rate": 9.877231524729078e-06, | |
| "loss": 0.7472, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 320534.59375, | |
| "learning_rate": 9.600310151937527e-06, | |
| "loss": 0.7602, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 295994.96875, | |
| "learning_rate": 9.323388779145974e-06, | |
| "loss": 0.7368, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 221153.6875, | |
| "learning_rate": 9.046467406354423e-06, | |
| "loss": 0.7492, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 250322.6875, | |
| "learning_rate": 8.76954603356287e-06, | |
| "loss": 0.7582, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 364772.46875, | |
| "learning_rate": 8.492624660771318e-06, | |
| "loss": 0.747, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 226077.453125, | |
| "learning_rate": 8.215703287979767e-06, | |
| "loss": 0.7515, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 260477.265625, | |
| "learning_rate": 7.938781915188214e-06, | |
| "loss": 0.7402, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 380977.46875, | |
| "learning_rate": 7.661860542396663e-06, | |
| "loss": 0.7512, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 1052145.875, | |
| "learning_rate": 7.38493916960511e-06, | |
| "loss": 0.7589, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 213601.171875, | |
| "learning_rate": 7.1080177968135575e-06, | |
| "loss": 0.7581, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 467877.71875, | |
| "learning_rate": 6.831096424022006e-06, | |
| "loss": 0.7464, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 566204.3125, | |
| "learning_rate": 6.554175051230454e-06, | |
| "loss": 0.7557, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 330177.5625, | |
| "learning_rate": 6.277253678438902e-06, | |
| "loss": 0.7304, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 319206.65625, | |
| "learning_rate": 6.00033230564735e-06, | |
| "loss": 0.736, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 410904.90625, | |
| "learning_rate": 5.723410932855798e-06, | |
| "loss": 0.752, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 375357.15625, | |
| "learning_rate": 5.4464895600642454e-06, | |
| "loss": 0.7358, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 307483.375, | |
| "learning_rate": 5.169568187272693e-06, | |
| "loss": 0.7749, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 434196.59375, | |
| "learning_rate": 4.892646814481141e-06, | |
| "loss": 0.7324, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 554119.4375, | |
| "learning_rate": 4.61572544168959e-06, | |
| "loss": 0.7096, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 224006.78125, | |
| "learning_rate": 4.338804068898038e-06, | |
| "loss": 0.7205, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 215164.78125, | |
| "learning_rate": 4.0618826961064855e-06, | |
| "loss": 0.732, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 391820.53125, | |
| "learning_rate": 3.7849613233149334e-06, | |
| "loss": 0.7228, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 306694.15625, | |
| "learning_rate": 3.5080399505233816e-06, | |
| "loss": 0.7507, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 467995.03125, | |
| "learning_rate": 3.2311185777318295e-06, | |
| "loss": 0.7433, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 284427.625, | |
| "learning_rate": 2.9541972049402773e-06, | |
| "loss": 0.7047, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 193486.921875, | |
| "learning_rate": 2.6772758321487256e-06, | |
| "loss": 0.7646, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 262856.15625, | |
| "learning_rate": 2.4003544593571735e-06, | |
| "loss": 0.7306, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 267263.5, | |
| "learning_rate": 2.1234330865656213e-06, | |
| "loss": 0.7631, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 229738.515625, | |
| "learning_rate": 1.8465117137740692e-06, | |
| "loss": 0.731, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 528997.6875, | |
| "learning_rate": 1.569590340982517e-06, | |
| "loss": 0.7393, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 346399.28125, | |
| "learning_rate": 1.292668968190965e-06, | |
| "loss": 0.7304, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 315753.28125, | |
| "learning_rate": 1.015747595399413e-06, | |
| "loss": 0.7239, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 245733.109375, | |
| "learning_rate": 7.388262226078609e-07, | |
| "loss": 0.7067, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 423298.15625, | |
| "learning_rate": 4.619048498163089e-07, | |
| "loss": 0.734, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 361011.8125, | |
| "learning_rate": 1.8498347702475678e-07, | |
| "loss": 0.7356, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.68, | |
| "eval_loss": 0.750106155872345, | |
| "eval_runtime": 570.5256, | |
| "eval_samples_per_second": 17.528, | |
| "eval_steps_per_second": 2.922, | |
| "step": 108334 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 108334, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "total_flos": 2.328341808688051e+18, | |
| "train_batch_size": 6, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |