| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9992652461425422, | |
| "eval_steps": 250, | |
| "global_step": 680, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0014695077149155032, | |
| "grad_norm": 11.337424189328551, | |
| "learning_rate": 4.7619047619047613e-08, | |
| "loss": 1.5474, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0029390154298310064, | |
| "grad_norm": 11.09462136805898, | |
| "learning_rate": 9.523809523809523e-08, | |
| "loss": 1.5705, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.00440852314474651, | |
| "grad_norm": 11.058506611133362, | |
| "learning_rate": 1.4285714285714285e-07, | |
| "loss": 1.4847, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.005878030859662013, | |
| "grad_norm": 10.9831239658005, | |
| "learning_rate": 1.9047619047619045e-07, | |
| "loss": 1.5536, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0073475385745775165, | |
| "grad_norm": 10.8860240443791, | |
| "learning_rate": 2.3809523809523806e-07, | |
| "loss": 1.5523, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.00881704628949302, | |
| "grad_norm": 12.310500944697788, | |
| "learning_rate": 2.857142857142857e-07, | |
| "loss": 1.5543, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.010286554004408524, | |
| "grad_norm": 10.339190259363582, | |
| "learning_rate": 3.333333333333333e-07, | |
| "loss": 1.5789, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.011756061719324026, | |
| "grad_norm": 11.75096794982965, | |
| "learning_rate": 3.809523809523809e-07, | |
| "loss": 1.4619, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.01322556943423953, | |
| "grad_norm": 10.983629821806812, | |
| "learning_rate": 4.285714285714285e-07, | |
| "loss": 1.5461, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.014695077149155033, | |
| "grad_norm": 10.026453006105893, | |
| "learning_rate": 4.761904761904761e-07, | |
| "loss": 1.594, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.016164584864070537, | |
| "grad_norm": 9.2515708413017, | |
| "learning_rate": 5.238095238095238e-07, | |
| "loss": 1.4904, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.01763409257898604, | |
| "grad_norm": 8.8513471084721, | |
| "learning_rate": 5.714285714285714e-07, | |
| "loss": 1.5626, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.019103600293901544, | |
| "grad_norm": 9.013166716897066, | |
| "learning_rate": 6.19047619047619e-07, | |
| "loss": 1.5522, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.020573108008817047, | |
| "grad_norm": 8.149687442441557, | |
| "learning_rate": 6.666666666666666e-07, | |
| "loss": 1.5075, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.02204261572373255, | |
| "grad_norm": 6.880717564693127, | |
| "learning_rate": 7.142857142857143e-07, | |
| "loss": 1.538, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.02351212343864805, | |
| "grad_norm": 6.545463429781402, | |
| "learning_rate": 7.619047619047618e-07, | |
| "loss": 1.5044, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.024981631153563555, | |
| "grad_norm": 6.093488294374897, | |
| "learning_rate": 8.095238095238095e-07, | |
| "loss": 1.4779, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.02645113886847906, | |
| "grad_norm": 6.4020806034329425, | |
| "learning_rate": 8.57142857142857e-07, | |
| "loss": 1.4431, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.027920646583394562, | |
| "grad_norm": 6.142908310782321, | |
| "learning_rate": 9.047619047619047e-07, | |
| "loss": 1.5965, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.029390154298310066, | |
| "grad_norm": 5.136583641076265, | |
| "learning_rate": 9.523809523809522e-07, | |
| "loss": 1.3968, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.03085966201322557, | |
| "grad_norm": 5.02265599787556, | |
| "learning_rate": 1e-06, | |
| "loss": 1.4007, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.03232916972814107, | |
| "grad_norm": 4.976860924734495, | |
| "learning_rate": 9.999943184333936e-07, | |
| "loss": 1.4378, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.03379867744305658, | |
| "grad_norm": 4.89918762254008, | |
| "learning_rate": 9.999772738626954e-07, | |
| "loss": 1.4341, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.03526818515797208, | |
| "grad_norm": 4.659631180053279, | |
| "learning_rate": 9.999488666752648e-07, | |
| "loss": 1.4178, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.036737692872887584, | |
| "grad_norm": 4.542979691946772, | |
| "learning_rate": 9.99909097516691e-07, | |
| "loss": 1.4933, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.03820720058780309, | |
| "grad_norm": 4.342835973040556, | |
| "learning_rate": 9.998579672907788e-07, | |
| "loss": 1.4417, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.03967670830271859, | |
| "grad_norm": 3.2871765127170853, | |
| "learning_rate": 9.99795477159527e-07, | |
| "loss": 1.423, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.041146216017634095, | |
| "grad_norm": 3.2008427389003766, | |
| "learning_rate": 9.99721628543103e-07, | |
| "loss": 1.3282, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.0426157237325496, | |
| "grad_norm": 3.1114560975594836, | |
| "learning_rate": 9.996364231198103e-07, | |
| "loss": 1.3891, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.0440852314474651, | |
| "grad_norm": 2.9307530666626884, | |
| "learning_rate": 9.9953986282605e-07, | |
| "loss": 1.3589, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.045554739162380606, | |
| "grad_norm": 2.9544258327932, | |
| "learning_rate": 9.99431949856277e-07, | |
| "loss": 1.3487, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.0470242468772961, | |
| "grad_norm": 2.9071124300801565, | |
| "learning_rate": 9.9931268666295e-07, | |
| "loss": 1.3462, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.048493754592211606, | |
| "grad_norm": 2.673673317642893, | |
| "learning_rate": 9.991820759564766e-07, | |
| "loss": 1.3972, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.04996326230712711, | |
| "grad_norm": 2.746698468930808, | |
| "learning_rate": 9.990401207051504e-07, | |
| "loss": 1.2998, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.05143277002204261, | |
| "grad_norm": 2.699798149588662, | |
| "learning_rate": 9.98886824135084e-07, | |
| "loss": 1.3451, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.05290227773695812, | |
| "grad_norm": 2.7302270842982987, | |
| "learning_rate": 9.98722189730136e-07, | |
| "loss": 1.392, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.05437178545187362, | |
| "grad_norm": 2.535544436400141, | |
| "learning_rate": 9.985462212318322e-07, | |
| "loss": 1.3673, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.055841293166789124, | |
| "grad_norm": 2.5536809032653145, | |
| "learning_rate": 9.983589226392792e-07, | |
| "loss": 1.2696, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.05731080088170463, | |
| "grad_norm": 2.438242871155347, | |
| "learning_rate": 9.98160298209075e-07, | |
| "loss": 1.3818, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.05878030859662013, | |
| "grad_norm": 2.2111464394984814, | |
| "learning_rate": 9.97950352455211e-07, | |
| "loss": 1.2618, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.060249816311535635, | |
| "grad_norm": 2.0951995475252856, | |
| "learning_rate": 9.977290901489707e-07, | |
| "loss": 1.2963, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.06171932402645114, | |
| "grad_norm": 2.1449537100989846, | |
| "learning_rate": 9.9749651631882e-07, | |
| "loss": 1.2525, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.06318883174136664, | |
| "grad_norm": 2.617528852412968, | |
| "learning_rate": 9.972526362502937e-07, | |
| "loss": 1.3144, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.06465833945628215, | |
| "grad_norm": 2.210234169368085, | |
| "learning_rate": 9.969974554858754e-07, | |
| "loss": 1.2379, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.06612784717119764, | |
| "grad_norm": 2.42457485573731, | |
| "learning_rate": 9.967309798248707e-07, | |
| "loss": 1.2573, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.06759735488611315, | |
| "grad_norm": 2.0254014234209405, | |
| "learning_rate": 9.96453215323277e-07, | |
| "loss": 1.2536, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.06906686260102865, | |
| "grad_norm": 2.0725386643920958, | |
| "learning_rate": 9.961641682936442e-07, | |
| "loss": 1.2687, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.07053637031594416, | |
| "grad_norm": 1.8563249903888095, | |
| "learning_rate": 9.95863845304932e-07, | |
| "loss": 1.2335, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.07200587803085966, | |
| "grad_norm": 1.9548577428039762, | |
| "learning_rate": 9.955522531823606e-07, | |
| "loss": 1.3056, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.07347538574577517, | |
| "grad_norm": 1.855842269546498, | |
| "learning_rate": 9.952293990072557e-07, | |
| "loss": 1.2671, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.07494489346069066, | |
| "grad_norm": 1.7972312223565392, | |
| "learning_rate": 9.948952901168874e-07, | |
| "loss": 1.2718, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.07641440117560618, | |
| "grad_norm": 1.6423712015458758, | |
| "learning_rate": 9.945499341043033e-07, | |
| "loss": 1.2043, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.07788390889052167, | |
| "grad_norm": 1.6028195930568658, | |
| "learning_rate": 9.94193338818156e-07, | |
| "loss": 1.2136, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.07935341660543718, | |
| "grad_norm": 1.5486505262354593, | |
| "learning_rate": 9.938255123625251e-07, | |
| "loss": 1.2556, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.08082292432035268, | |
| "grad_norm": 1.5343630961919048, | |
| "learning_rate": 9.934464630967328e-07, | |
| "loss": 1.2519, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.08229243203526819, | |
| "grad_norm": 1.662816938451208, | |
| "learning_rate": 9.930561996351533e-07, | |
| "loss": 1.2482, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.08376193975018369, | |
| "grad_norm": 1.5931246608007381, | |
| "learning_rate": 9.926547308470183e-07, | |
| "loss": 1.1625, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.0852314474650992, | |
| "grad_norm": 1.5447707955729764, | |
| "learning_rate": 9.922420658562144e-07, | |
| "loss": 1.1873, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.0867009551800147, | |
| "grad_norm": 1.644709907245824, | |
| "learning_rate": 9.91818214041076e-07, | |
| "loss": 1.1914, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.0881704628949302, | |
| "grad_norm": 1.6366535743931827, | |
| "learning_rate": 9.913831850341725e-07, | |
| "loss": 1.1871, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.0896399706098457, | |
| "grad_norm": 1.5571790849844254, | |
| "learning_rate": 9.90936988722089e-07, | |
| "loss": 1.2605, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.09110947832476121, | |
| "grad_norm": 3.7097540672301386, | |
| "learning_rate": 9.904796352452019e-07, | |
| "loss": 1.1212, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.09257898603967671, | |
| "grad_norm": 1.5517830620847208, | |
| "learning_rate": 9.900111349974478e-07, | |
| "loss": 1.1623, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.0940484937545922, | |
| "grad_norm": 1.5680093555121737, | |
| "learning_rate": 9.895314986260886e-07, | |
| "loss": 1.1552, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.09551800146950772, | |
| "grad_norm": 1.6209106318425315, | |
| "learning_rate": 9.890407370314677e-07, | |
| "loss": 1.1337, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.09698750918442321, | |
| "grad_norm": 1.4100673270240494, | |
| "learning_rate": 9.885388613667644e-07, | |
| "loss": 1.2156, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.09845701689933872, | |
| "grad_norm": 1.5012099159985126, | |
| "learning_rate": 9.880258830377386e-07, | |
| "loss": 1.2427, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.09992652461425422, | |
| "grad_norm": 1.5660607972935188, | |
| "learning_rate": 9.875018137024721e-07, | |
| "loss": 1.1729, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.10139603232916973, | |
| "grad_norm": 1.5666357672603266, | |
| "learning_rate": 9.869666652711049e-07, | |
| "loss": 1.2838, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.10286554004408523, | |
| "grad_norm": 1.5178831815539953, | |
| "learning_rate": 9.864204499055622e-07, | |
| "loss": 1.2145, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.10433504775900074, | |
| "grad_norm": 1.71647656199295, | |
| "learning_rate": 9.858631800192804e-07, | |
| "loss": 1.1724, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.10580455547391623, | |
| "grad_norm": 1.5138224523065718, | |
| "learning_rate": 9.852948682769234e-07, | |
| "loss": 1.2672, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.10727406318883174, | |
| "grad_norm": 1.5726438604764952, | |
| "learning_rate": 9.84715527594095e-07, | |
| "loss": 1.1505, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.10874357090374724, | |
| "grad_norm": 1.5196898825337892, | |
| "learning_rate": 9.841251711370457e-07, | |
| "loss": 1.1801, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.11021307861866275, | |
| "grad_norm": 1.5385492133663783, | |
| "learning_rate": 9.83523812322374e-07, | |
| "loss": 1.2523, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.11168258633357825, | |
| "grad_norm": 1.4506696813615474, | |
| "learning_rate": 9.829114648167206e-07, | |
| "loss": 1.1798, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.11315209404849376, | |
| "grad_norm": 1.5471986855965765, | |
| "learning_rate": 9.822881425364578e-07, | |
| "loss": 1.1227, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.11462160176340926, | |
| "grad_norm": 1.4810764356218553, | |
| "learning_rate": 9.81653859647374e-07, | |
| "loss": 1.1904, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.11609110947832477, | |
| "grad_norm": 1.458422270393885, | |
| "learning_rate": 9.810086305643511e-07, | |
| "loss": 1.1688, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.11756061719324026, | |
| "grad_norm": 1.4150019691387836, | |
| "learning_rate": 9.803524699510372e-07, | |
| "loss": 1.1679, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.11903012490815577, | |
| "grad_norm": 1.5594916312404665, | |
| "learning_rate": 9.79685392719513e-07, | |
| "loss": 1.1915, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.12049963262307127, | |
| "grad_norm": 1.5378772800136011, | |
| "learning_rate": 9.790074140299535e-07, | |
| "loss": 1.2446, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.12196914033798678, | |
| "grad_norm": 1.3745260542581244, | |
| "learning_rate": 9.783185492902831e-07, | |
| "loss": 1.1894, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.12343864805290228, | |
| "grad_norm": 1.4281455021507763, | |
| "learning_rate": 9.776188141558253e-07, | |
| "loss": 1.1535, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.12490815576781777, | |
| "grad_norm": 1.4827991467262824, | |
| "learning_rate": 9.769082245289472e-07, | |
| "loss": 1.1812, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.12637766348273327, | |
| "grad_norm": 1.4091217681303907, | |
| "learning_rate": 9.76186796558698e-07, | |
| "loss": 1.1733, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.1278471711976488, | |
| "grad_norm": 1.504340633323682, | |
| "learning_rate": 9.754545466404423e-07, | |
| "loss": 1.1891, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.1293166789125643, | |
| "grad_norm": 1.5341255256297364, | |
| "learning_rate": 9.747114914154862e-07, | |
| "loss": 1.1941, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.1307861866274798, | |
| "grad_norm": 1.448718524801724, | |
| "learning_rate": 9.73957647770701e-07, | |
| "loss": 1.2118, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.13225569434239529, | |
| "grad_norm": 1.4171871543912984, | |
| "learning_rate": 9.731930328381384e-07, | |
| "loss": 1.1745, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.1337252020573108, | |
| "grad_norm": 1.487952395310904, | |
| "learning_rate": 9.72417663994641e-07, | |
| "loss": 1.1625, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.1351947097722263, | |
| "grad_norm": 1.5914017741584792, | |
| "learning_rate": 9.716315588614472e-07, | |
| "loss": 1.1051, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.1366642174871418, | |
| "grad_norm": 1.6391622049937462, | |
| "learning_rate": 9.708347353037924e-07, | |
| "loss": 1.143, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.1381337252020573, | |
| "grad_norm": 1.4437878891336424, | |
| "learning_rate": 9.700272114305008e-07, | |
| "loss": 1.1648, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.13960323291697282, | |
| "grad_norm": 1.4177630579747587, | |
| "learning_rate": 9.69209005593575e-07, | |
| "loss": 1.1999, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.14107274063188832, | |
| "grad_norm": 1.5120811492994586, | |
| "learning_rate": 9.68380136387779e-07, | |
| "loss": 1.2242, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.14254224834680382, | |
| "grad_norm": 1.4738631960216506, | |
| "learning_rate": 9.67540622650215e-07, | |
| "loss": 1.1548, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.14401175606171931, | |
| "grad_norm": 1.5620452244313792, | |
| "learning_rate": 9.66690483459896e-07, | |
| "loss": 1.1875, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.14548126377663484, | |
| "grad_norm": 1.5673649210031055, | |
| "learning_rate": 9.658297381373117e-07, | |
| "loss": 1.198, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.14695077149155034, | |
| "grad_norm": 1.4854078674557176, | |
| "learning_rate": 9.649584062439897e-07, | |
| "loss": 1.2242, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.14842027920646583, | |
| "grad_norm": 1.5887802038965162, | |
| "learning_rate": 9.640765075820508e-07, | |
| "loss": 1.23, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.14988978692138133, | |
| "grad_norm": 1.5956321149777724, | |
| "learning_rate": 9.631840621937585e-07, | |
| "loss": 1.226, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.15135929463629685, | |
| "grad_norm": 1.4446949899977264, | |
| "learning_rate": 9.622810903610653e-07, | |
| "loss": 1.1796, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.15282880235121235, | |
| "grad_norm": 1.4023788130375525, | |
| "learning_rate": 9.613676126051488e-07, | |
| "loss": 1.1589, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.15429831006612785, | |
| "grad_norm": 1.5491829383620404, | |
| "learning_rate": 9.604436496859482e-07, | |
| "loss": 1.237, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.15576781778104334, | |
| "grad_norm": 1.3527676295882165, | |
| "learning_rate": 9.595092226016912e-07, | |
| "loss": 1.1329, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.15723732549595884, | |
| "grad_norm": 1.5174098623642012, | |
| "learning_rate": 9.585643525884163e-07, | |
| "loss": 1.179, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.15870683321087437, | |
| "grad_norm": 1.3787786506819375, | |
| "learning_rate": 9.576090611194915e-07, | |
| "loss": 1.1199, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.16017634092578986, | |
| "grad_norm": 1.4679013393236473, | |
| "learning_rate": 9.566433699051248e-07, | |
| "loss": 1.2768, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.16164584864070536, | |
| "grad_norm": 1.4363459788248092, | |
| "learning_rate": 9.556673008918725e-07, | |
| "loss": 1.1434, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.16311535635562086, | |
| "grad_norm": 1.5373366464111682, | |
| "learning_rate": 9.546808762621385e-07, | |
| "loss": 1.1391, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.16458486407053638, | |
| "grad_norm": 1.397660653092303, | |
| "learning_rate": 9.536841184336725e-07, | |
| "loss": 1.1599, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.16605437178545188, | |
| "grad_norm": 1.5338825854004605, | |
| "learning_rate": 9.526770500590576e-07, | |
| "loss": 1.0599, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.16752387950036737, | |
| "grad_norm": 1.7726346336471452, | |
| "learning_rate": 9.516596940251986e-07, | |
| "loss": 1.1347, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.16899338721528287, | |
| "grad_norm": 1.401274657699429, | |
| "learning_rate": 9.506320734527997e-07, | |
| "loss": 1.1538, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.1704628949301984, | |
| "grad_norm": 1.4313194549405384, | |
| "learning_rate": 9.495942116958395e-07, | |
| "loss": 1.2409, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.1719324026451139, | |
| "grad_norm": 1.338422913627801, | |
| "learning_rate": 9.485461323410411e-07, | |
| "loss": 1.2274, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.1734019103600294, | |
| "grad_norm": 1.414029707516539, | |
| "learning_rate": 9.474878592073352e-07, | |
| "loss": 1.1813, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.17487141807494488, | |
| "grad_norm": 1.5090772581920886, | |
| "learning_rate": 9.464194163453188e-07, | |
| "loss": 1.1576, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.1763409257898604, | |
| "grad_norm": 1.397279646873424, | |
| "learning_rate": 9.45340828036709e-07, | |
| "loss": 1.1172, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.1778104335047759, | |
| "grad_norm": 1.4335773048462084, | |
| "learning_rate": 9.442521187937911e-07, | |
| "loss": 1.2413, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.1792799412196914, | |
| "grad_norm": 1.4697864318337606, | |
| "learning_rate": 9.431533133588616e-07, | |
| "loss": 1.1551, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.1807494489346069, | |
| "grad_norm": 1.418392328936289, | |
| "learning_rate": 9.420444367036653e-07, | |
| "loss": 1.1411, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.18221895664952242, | |
| "grad_norm": 1.3863098352226801, | |
| "learning_rate": 9.409255140288288e-07, | |
| "loss": 1.1388, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.18368846436443792, | |
| "grad_norm": 1.4189204374991593, | |
| "learning_rate": 9.397965707632866e-07, | |
| "loss": 1.1512, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.18515797207935342, | |
| "grad_norm": 1.524052045769854, | |
| "learning_rate": 9.386576325637043e-07, | |
| "loss": 1.1759, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.1866274797942689, | |
| "grad_norm": 1.405091479549588, | |
| "learning_rate": 9.375087253138951e-07, | |
| "loss": 1.166, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.1880969875091844, | |
| "grad_norm": 1.544876861673977, | |
| "learning_rate": 9.363498751242307e-07, | |
| "loss": 1.1309, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.18956649522409993, | |
| "grad_norm": 1.423691492612366, | |
| "learning_rate": 9.351811083310497e-07, | |
| "loss": 1.1394, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.19103600293901543, | |
| "grad_norm": 1.466545968863769, | |
| "learning_rate": 9.340024514960574e-07, | |
| "loss": 1.1772, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.19250551065393093, | |
| "grad_norm": 1.4099417566779877, | |
| "learning_rate": 9.328139314057233e-07, | |
| "loss": 1.116, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.19397501836884642, | |
| "grad_norm": 1.3751112771189664, | |
| "learning_rate": 9.316155750706713e-07, | |
| "loss": 1.1249, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.19544452608376195, | |
| "grad_norm": 1.440365523396416, | |
| "learning_rate": 9.304074097250671e-07, | |
| "loss": 1.1607, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.19691403379867745, | |
| "grad_norm": 1.3429807064760693, | |
| "learning_rate": 9.291894628259979e-07, | |
| "loss": 1.148, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.19838354151359294, | |
| "grad_norm": 1.6100321463659257, | |
| "learning_rate": 9.279617620528496e-07, | |
| "loss": 1.17, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.19985304922850844, | |
| "grad_norm": 1.5750772736599945, | |
| "learning_rate": 9.26724335306677e-07, | |
| "loss": 1.1625, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.20132255694342396, | |
| "grad_norm": 1.4373009264887187, | |
| "learning_rate": 9.254772107095701e-07, | |
| "loss": 1.1595, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.20279206465833946, | |
| "grad_norm": 1.3522430478527927, | |
| "learning_rate": 9.242204166040148e-07, | |
| "loss": 1.105, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.20426157237325496, | |
| "grad_norm": 1.496220468060292, | |
| "learning_rate": 9.229539815522485e-07, | |
| "loss": 1.1729, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.20573108008817045, | |
| "grad_norm": 1.3692383152400416, | |
| "learning_rate": 9.216779343356119e-07, | |
| "loss": 1.0892, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.20720058780308598, | |
| "grad_norm": 1.415058934382544, | |
| "learning_rate": 9.203923039538939e-07, | |
| "loss": 1.098, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.20867009551800147, | |
| "grad_norm": 1.3906133545085528, | |
| "learning_rate": 9.190971196246731e-07, | |
| "loss": 1.1862, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.21013960323291697, | |
| "grad_norm": 1.4853242615363982, | |
| "learning_rate": 9.177924107826535e-07, | |
| "loss": 1.125, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.21160911094783247, | |
| "grad_norm": 1.4673768423852198, | |
| "learning_rate": 9.164782070789961e-07, | |
| "loss": 1.2024, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.213078618662748, | |
| "grad_norm": 1.357361299564151, | |
| "learning_rate": 9.151545383806441e-07, | |
| "loss": 1.0744, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.2145481263776635, | |
| "grad_norm": 1.4090842422695524, | |
| "learning_rate": 9.138214347696453e-07, | |
| "loss": 1.2193, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.216017634092579, | |
| "grad_norm": 1.403649440383755, | |
| "learning_rate": 9.124789265424674e-07, | |
| "loss": 1.1808, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.21748714180749448, | |
| "grad_norm": 1.3566301062411306, | |
| "learning_rate": 9.1112704420931e-07, | |
| "loss": 1.1419, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.21895664952240998, | |
| "grad_norm": 1.3695324273141942, | |
| "learning_rate": 9.097658184934114e-07, | |
| "loss": 1.1495, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.2204261572373255, | |
| "grad_norm": 1.4138005238611118, | |
| "learning_rate": 9.083952803303496e-07, | |
| "loss": 1.2471, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.221895664952241, | |
| "grad_norm": 1.3414103194314155, | |
| "learning_rate": 9.070154608673402e-07, | |
| "loss": 1.1672, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.2233651726671565, | |
| "grad_norm": 1.438725688375383, | |
| "learning_rate": 9.056263914625277e-07, | |
| "loss": 1.2144, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.224834680382072, | |
| "grad_norm": 1.564966532077944, | |
| "learning_rate": 9.042281036842739e-07, | |
| "loss": 1.1981, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.22630418809698752, | |
| "grad_norm": 1.275239691393889, | |
| "learning_rate": 9.028206293104391e-07, | |
| "loss": 1.1098, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.22777369581190302, | |
| "grad_norm": 1.4419905057143438, | |
| "learning_rate": 9.014040003276611e-07, | |
| "loss": 1.1783, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.2292432035268185, | |
| "grad_norm": 1.4105436779658675, | |
| "learning_rate": 8.999782489306271e-07, | |
| "loss": 1.1513, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.230712711241734, | |
| "grad_norm": 1.4197672181122762, | |
| "learning_rate": 8.985434075213439e-07, | |
| "loss": 1.1433, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.23218221895664953, | |
| "grad_norm": 1.3670907633233818, | |
| "learning_rate": 8.970995087083992e-07, | |
| "loss": 1.03, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.23365172667156503, | |
| "grad_norm": 1.3603227901129349, | |
| "learning_rate": 8.956465853062222e-07, | |
| "loss": 1.0851, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.23512123438648053, | |
| "grad_norm": 1.3852573758765583, | |
| "learning_rate": 8.941846703343372e-07, | |
| "loss": 1.2005, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.23659074210139602, | |
| "grad_norm": 1.3638595388577004, | |
| "learning_rate": 8.927137970166135e-07, | |
| "loss": 1.1217, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.23806024981631155, | |
| "grad_norm": 1.4806310433707142, | |
| "learning_rate": 8.912339987805099e-07, | |
| "loss": 1.1339, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.23952975753122704, | |
| "grad_norm": 1.4400386045554132, | |
| "learning_rate": 8.897453092563153e-07, | |
| "loss": 1.1343, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.24099926524614254, | |
| "grad_norm": 1.3697227632539914, | |
| "learning_rate": 8.882477622763846e-07, | |
| "loss": 1.1529, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.24246877296105804, | |
| "grad_norm": 1.3503406924429706, | |
| "learning_rate": 8.867413918743693e-07, | |
| "loss": 1.164, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.24393828067597356, | |
| "grad_norm": 1.3850072048890962, | |
| "learning_rate": 8.852262322844444e-07, | |
| "loss": 1.153, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.24540778839088906, | |
| "grad_norm": 1.359563698104864, | |
| "learning_rate": 8.837023179405308e-07, | |
| "loss": 1.1724, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.24687729610580456, | |
| "grad_norm": 1.3978149085145892, | |
| "learning_rate": 8.821696834755117e-07, | |
| "loss": 1.0965, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.24834680382072005, | |
| "grad_norm": 1.404265065477685, | |
| "learning_rate": 8.806283637204462e-07, | |
| "loss": 1.1879, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.24981631153563555, | |
| "grad_norm": 1.403937125406496, | |
| "learning_rate": 8.790783937037776e-07, | |
| "loss": 1.1251, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.25128581925055105, | |
| "grad_norm": 1.4373795084844492, | |
| "learning_rate": 8.775198086505375e-07, | |
| "loss": 1.1347, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.25275532696546654, | |
| "grad_norm": 1.4706572304655665, | |
| "learning_rate": 8.759526439815455e-07, | |
| "loss": 1.0477, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.2542248346803821, | |
| "grad_norm": 1.4008092483314147, | |
| "learning_rate": 8.743769353126029e-07, | |
| "loss": 1.1761, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.2556943423952976, | |
| "grad_norm": 1.3764920528768414, | |
| "learning_rate": 8.727927184536849e-07, | |
| "loss": 1.1277, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.2571638501102131, | |
| "grad_norm": 1.4213614638860257, | |
| "learning_rate": 8.712000294081259e-07, | |
| "loss": 1.0911, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.2586333578251286, | |
| "grad_norm": 1.3408094274422444, | |
| "learning_rate": 8.695989043718015e-07, | |
| "loss": 1.1602, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.2601028655400441, | |
| "grad_norm": 1.4025428044724146, | |
| "learning_rate": 8.679893797323058e-07, | |
| "loss": 1.1325, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.2615723732549596, | |
| "grad_norm": 1.3560427968027609, | |
| "learning_rate": 8.663714920681245e-07, | |
| "loss": 1.1789, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.2630418809698751, | |
| "grad_norm": 1.4997523461667033, | |
| "learning_rate": 8.64745278147804e-07, | |
| "loss": 1.1578, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.26451138868479057, | |
| "grad_norm": 1.4091572819165725, | |
| "learning_rate": 8.631107749291148e-07, | |
| "loss": 1.1344, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.2659808963997061, | |
| "grad_norm": 1.5029746752542426, | |
| "learning_rate": 8.614680195582127e-07, | |
| "loss": 1.1829, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.2674504041146216, | |
| "grad_norm": 1.4231917336486979, | |
| "learning_rate": 8.598170493687939e-07, | |
| "loss": 1.097, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.2689199118295371, | |
| "grad_norm": 1.3738447004488614, | |
| "learning_rate": 8.581579018812468e-07, | |
| "loss": 1.1539, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.2703894195444526, | |
| "grad_norm": 1.4706711617380483, | |
| "learning_rate": 8.564906148017992e-07, | |
| "loss": 1.179, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.2718589272593681, | |
| "grad_norm": 1.3788046176437054, | |
| "learning_rate": 8.548152260216613e-07, | |
| "loss": 1.1003, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.2733284349742836, | |
| "grad_norm": 1.3707247735299555, | |
| "learning_rate": 8.531317736161652e-07, | |
| "loss": 1.0604, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.2747979426891991, | |
| "grad_norm": 1.3354127082342266, | |
| "learning_rate": 8.514402958438987e-07, | |
| "loss": 1.1121, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.2762674504041146, | |
| "grad_norm": 1.336615144205344, | |
| "learning_rate": 8.497408311458362e-07, | |
| "loss": 1.0885, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.2777369581190301, | |
| "grad_norm": 1.3645293415775273, | |
| "learning_rate": 8.480334181444652e-07, | |
| "loss": 1.1334, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.27920646583394565, | |
| "grad_norm": 1.405142088376881, | |
| "learning_rate": 8.463180956429085e-07, | |
| "loss": 1.1685, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.28067597354886115, | |
| "grad_norm": 1.3868866326557587, | |
| "learning_rate": 8.445949026240424e-07, | |
| "loss": 1.2174, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.28214548126377664, | |
| "grad_norm": 1.335338615058983, | |
| "learning_rate": 8.428638782496105e-07, | |
| "loss": 1.1235, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.28361498897869214, | |
| "grad_norm": 1.3713807109243261, | |
| "learning_rate": 8.411250618593337e-07, | |
| "loss": 1.2191, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.28508449669360764, | |
| "grad_norm": 1.41726090350252, | |
| "learning_rate": 8.393784929700169e-07, | |
| "loss": 1.1522, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.28655400440852313, | |
| "grad_norm": 1.3529238461119495, | |
| "learning_rate": 8.376242112746499e-07, | |
| "loss": 1.0376, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.28802351212343863, | |
| "grad_norm": 1.4100766083722753, | |
| "learning_rate": 8.358622566415057e-07, | |
| "loss": 1.0645, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.2894930198383541, | |
| "grad_norm": 1.3130064224637852, | |
| "learning_rate": 8.340926691132348e-07, | |
| "loss": 1.1029, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.2909625275532697, | |
| "grad_norm": 1.3572153763312709, | |
| "learning_rate": 8.323154889059549e-07, | |
| "loss": 1.1309, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.2924320352681852, | |
| "grad_norm": 1.542624539487457, | |
| "learning_rate": 8.305307564083368e-07, | |
| "loss": 1.2663, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.29390154298310067, | |
| "grad_norm": 1.3525660807398292, | |
| "learning_rate": 8.287385121806869e-07, | |
| "loss": 1.1509, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.29537105069801617, | |
| "grad_norm": 1.514167322393567, | |
| "learning_rate": 8.26938796954025e-07, | |
| "loss": 1.233, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.29684055841293167, | |
| "grad_norm": 1.3922047754873006, | |
| "learning_rate": 8.251316516291586e-07, | |
| "loss": 1.1035, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.29831006612784716, | |
| "grad_norm": 1.450031885451194, | |
| "learning_rate": 8.233171172757539e-07, | |
| "loss": 1.1351, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.29977957384276266, | |
| "grad_norm": 1.3679188570513114, | |
| "learning_rate": 8.214952351314022e-07, | |
| "loss": 1.105, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.30124908155767816, | |
| "grad_norm": 1.3734569453159238, | |
| "learning_rate": 8.196660466006823e-07, | |
| "loss": 1.1665, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.3027185892725937, | |
| "grad_norm": 1.4252740827310968, | |
| "learning_rate": 8.178295932542205e-07, | |
| "loss": 1.1422, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.3041880969875092, | |
| "grad_norm": 1.38166857353608, | |
| "learning_rate": 8.159859168277444e-07, | |
| "loss": 1.0333, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.3056576047024247, | |
| "grad_norm": 1.4722910903729798, | |
| "learning_rate": 8.141350592211358e-07, | |
| "loss": 1.134, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.3071271124173402, | |
| "grad_norm": 1.4074834904053308, | |
| "learning_rate": 8.122770624974778e-07, | |
| "loss": 1.0712, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.3085966201322557, | |
| "grad_norm": 1.4280613261841697, | |
| "learning_rate": 8.10411968882099e-07, | |
| "loss": 1.1277, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.3100661278471712, | |
| "grad_norm": 1.3743456352090058, | |
| "learning_rate": 8.085398207616138e-07, | |
| "loss": 1.2189, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.3115356355620867, | |
| "grad_norm": 1.4591835041502073, | |
| "learning_rate": 8.06660660682959e-07, | |
| "loss": 1.0779, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.3130051432770022, | |
| "grad_norm": 1.3896625615186315, | |
| "learning_rate": 8.047745313524275e-07, | |
| "loss": 1.118, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.3144746509919177, | |
| "grad_norm": 1.4095222295795578, | |
| "learning_rate": 8.028814756346967e-07, | |
| "loss": 1.1015, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.31594415870683323, | |
| "grad_norm": 1.3835385844364931, | |
| "learning_rate": 8.009815365518554e-07, | |
| "loss": 1.1964, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.31741366642174873, | |
| "grad_norm": 1.4442582574354577, | |
| "learning_rate": 7.990747572824253e-07, | |
| "loss": 1.0994, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.3188831741366642, | |
| "grad_norm": 1.540269470315885, | |
| "learning_rate": 7.971611811603803e-07, | |
| "loss": 1.1612, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.3203526818515797, | |
| "grad_norm": 1.5012905943987096, | |
| "learning_rate": 7.952408516741607e-07, | |
| "loss": 1.1212, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.3218221895664952, | |
| "grad_norm": 1.377730602489123, | |
| "learning_rate": 7.933138124656864e-07, | |
| "loss": 1.136, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.3232916972814107, | |
| "grad_norm": 1.4387714462293955, | |
| "learning_rate": 7.913801073293638e-07, | |
| "loss": 1.1662, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.3247612049963262, | |
| "grad_norm": 1.3416069488877052, | |
| "learning_rate": 7.894397802110908e-07, | |
| "loss": 1.1242, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.3262307127112417, | |
| "grad_norm": 1.3608424909383003, | |
| "learning_rate": 7.87492875207259e-07, | |
| "loss": 1.2513, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.32770022042615726, | |
| "grad_norm": 1.3266709078386894, | |
| "learning_rate": 7.855394365637495e-07, | |
| "loss": 1.1342, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.32916972814107276, | |
| "grad_norm": 1.3442846396883863, | |
| "learning_rate": 7.835795086749299e-07, | |
| "loss": 1.1801, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.33063923585598826, | |
| "grad_norm": 1.4096410255441036, | |
| "learning_rate": 7.816131360826434e-07, | |
| "loss": 1.1195, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.33210874357090375, | |
| "grad_norm": 1.3103219218263102, | |
| "learning_rate": 7.796403634751973e-07, | |
| "loss": 1.1207, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.33357825128581925, | |
| "grad_norm": 1.3677511713167119, | |
| "learning_rate": 7.776612356863477e-07, | |
| "loss": 1.0818, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.33504775900073475, | |
| "grad_norm": 1.4153862144691094, | |
| "learning_rate": 7.756757976942798e-07, | |
| "loss": 1.1155, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.33651726671565024, | |
| "grad_norm": 1.4438500129836211, | |
| "learning_rate": 7.736840946205865e-07, | |
| "loss": 1.193, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.33798677443056574, | |
| "grad_norm": 1.456870674708481, | |
| "learning_rate": 7.716861717292424e-07, | |
| "loss": 1.1199, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.33945628214548124, | |
| "grad_norm": 1.3923730055470556, | |
| "learning_rate": 7.696820744255756e-07, | |
| "loss": 1.2182, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.3409257898603968, | |
| "grad_norm": 1.370735300132156, | |
| "learning_rate": 7.676718482552353e-07, | |
| "loss": 1.134, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.3423952975753123, | |
| "grad_norm": 1.3480114555788523, | |
| "learning_rate": 7.65655538903157e-07, | |
| "loss": 1.0998, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.3438648052902278, | |
| "grad_norm": 1.4276815010976192, | |
| "learning_rate": 7.636331921925241e-07, | |
| "loss": 1.1696, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.3453343130051433, | |
| "grad_norm": 1.4170232593240883, | |
| "learning_rate": 7.61604854083727e-07, | |
| "loss": 1.116, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.3468038207200588, | |
| "grad_norm": 1.4134392902146415, | |
| "learning_rate": 7.595705706733178e-07, | |
| "loss": 1.1178, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.34827332843497427, | |
| "grad_norm": 1.3744159242762746, | |
| "learning_rate": 7.575303881929632e-07, | |
| "loss": 1.1824, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.34974283614988977, | |
| "grad_norm": 1.38324732435889, | |
| "learning_rate": 7.55484353008394e-07, | |
| "loss": 1.1422, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.35121234386480527, | |
| "grad_norm": 1.3881791765597504, | |
| "learning_rate": 7.534325116183508e-07, | |
| "loss": 1.1597, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.3526818515797208, | |
| "grad_norm": 1.396997073324397, | |
| "learning_rate": 7.513749106535278e-07, | |
| "loss": 1.1062, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.3541513592946363, | |
| "grad_norm": 1.443575506575812, | |
| "learning_rate": 7.493115968755125e-07, | |
| "loss": 1.1133, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.3556208670095518, | |
| "grad_norm": 1.4651722014658863, | |
| "learning_rate": 7.472426171757238e-07, | |
| "loss": 1.1609, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.3570903747244673, | |
| "grad_norm": 1.358202452421609, | |
| "learning_rate": 7.451680185743454e-07, | |
| "loss": 1.0965, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.3585598824393828, | |
| "grad_norm": 1.5175842365493015, | |
| "learning_rate": 7.430878482192579e-07, | |
| "loss": 1.0451, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.3600293901542983, | |
| "grad_norm": 1.4934774283507497, | |
| "learning_rate": 7.41002153384967e-07, | |
| "loss": 1.168, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.3614988978692138, | |
| "grad_norm": 1.4081881789418684, | |
| "learning_rate": 7.389109814715292e-07, | |
| "loss": 1.1344, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.3629684055841293, | |
| "grad_norm": 1.4765256904942174, | |
| "learning_rate": 7.368143800034745e-07, | |
| "loss": 1.1749, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.36443791329904485, | |
| "grad_norm": 1.3509670984266027, | |
| "learning_rate": 7.347123966287265e-07, | |
| "loss": 1.1334, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.36590742101396034, | |
| "grad_norm": 1.53782552923439, | |
| "learning_rate": 7.326050791175196e-07, | |
| "loss": 1.2062, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.36737692872887584, | |
| "grad_norm": 1.3428470092287972, | |
| "learning_rate": 7.304924753613127e-07, | |
| "loss": 1.0949, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.36737692872887584, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_loss": 2.756098508834839, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_runtime": 5.9542, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_samples_per_second": 15.619, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_steps_per_second": 2.015, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.36884643644379134, | |
| "grad_norm": 1.3548965420628396, | |
| "learning_rate": 7.283746333717014e-07, | |
| "loss": 1.1102, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.37031594415870683, | |
| "grad_norm": 1.3760628358821343, | |
| "learning_rate": 7.262516012793276e-07, | |
| "loss": 1.0957, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.37178545187362233, | |
| "grad_norm": 1.4363319701743325, | |
| "learning_rate": 7.241234273327838e-07, | |
| "loss": 1.0782, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.3732549595885378, | |
| "grad_norm": 1.3292955709874987, | |
| "learning_rate": 7.219901598975185e-07, | |
| "loss": 1.0904, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.3747244673034533, | |
| "grad_norm": 1.3562568095097107, | |
| "learning_rate": 7.198518474547354e-07, | |
| "loss": 1.0553, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.3761939750183688, | |
| "grad_norm": 1.3922411026579398, | |
| "learning_rate": 7.17708538600293e-07, | |
| "loss": 1.1265, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.3776634827332844, | |
| "grad_norm": 1.3917980021891811, | |
| "learning_rate": 7.155602820435992e-07, | |
| "loss": 1.153, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.37913299044819987, | |
| "grad_norm": 1.4243974023254953, | |
| "learning_rate": 7.134071266065051e-07, | |
| "loss": 1.1354, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.38060249816311537, | |
| "grad_norm": 1.4529979227761458, | |
| "learning_rate": 7.112491212221946e-07, | |
| "loss": 1.0746, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.38207200587803086, | |
| "grad_norm": 1.381064754681566, | |
| "learning_rate": 7.09086314934073e-07, | |
| "loss": 1.0771, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.38354151359294636, | |
| "grad_norm": 1.3494944459929241, | |
| "learning_rate": 7.069187568946524e-07, | |
| "loss": 1.1317, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.38501102130786186, | |
| "grad_norm": 1.367365128216842, | |
| "learning_rate": 7.047464963644342e-07, | |
| "loss": 1.0841, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.38648052902277735, | |
| "grad_norm": 1.4111229982907987, | |
| "learning_rate": 7.025695827107901e-07, | |
| "loss": 1.1456, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.38795003673769285, | |
| "grad_norm": 1.4469253951516285, | |
| "learning_rate": 7.003880654068395e-07, | |
| "loss": 1.1497, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.3894195444526084, | |
| "grad_norm": 1.438128214166176, | |
| "learning_rate": 6.98201994030326e-07, | |
| "loss": 1.1566, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.3908890521675239, | |
| "grad_norm": 1.392547852669767, | |
| "learning_rate": 6.960114182624902e-07, | |
| "loss": 1.1729, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.3923585598824394, | |
| "grad_norm": 1.401971190522276, | |
| "learning_rate": 6.938163878869405e-07, | |
| "loss": 1.0718, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.3938280675973549, | |
| "grad_norm": 1.396822074711564, | |
| "learning_rate": 6.916169527885221e-07, | |
| "loss": 1.1141, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.3952975753122704, | |
| "grad_norm": 1.3883010823518094, | |
| "learning_rate": 6.894131629521829e-07, | |
| "loss": 1.2072, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.3967670830271859, | |
| "grad_norm": 1.400367946344679, | |
| "learning_rate": 6.872050684618381e-07, | |
| "loss": 1.1512, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.3982365907421014, | |
| "grad_norm": 1.386986813686829, | |
| "learning_rate": 6.849927194992312e-07, | |
| "loss": 1.0693, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.3997060984570169, | |
| "grad_norm": 1.4393959811216241, | |
| "learning_rate": 6.827761663427943e-07, | |
| "loss": 1.1799, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.4011756061719324, | |
| "grad_norm": 1.3989135808754347, | |
| "learning_rate": 6.805554593665049e-07, | |
| "loss": 1.1858, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.4026451138868479, | |
| "grad_norm": 1.3294118191399562, | |
| "learning_rate": 6.783306490387414e-07, | |
| "loss": 1.051, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.4041146216017634, | |
| "grad_norm": 1.3425765430711993, | |
| "learning_rate": 6.761017859211359e-07, | |
| "loss": 1.0931, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.4055841293166789, | |
| "grad_norm": 1.3979046769171386, | |
| "learning_rate": 6.738689206674257e-07, | |
| "loss": 1.0426, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.4070536370315944, | |
| "grad_norm": 1.3494939507393635, | |
| "learning_rate": 6.716321040223014e-07, | |
| "loss": 1.0846, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.4085231447465099, | |
| "grad_norm": 1.38158252960459, | |
| "learning_rate": 6.693913868202539e-07, | |
| "loss": 1.0751, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.4099926524614254, | |
| "grad_norm": 1.3216203318442303, | |
| "learning_rate": 6.671468199844192e-07, | |
| "loss": 1.074, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.4114621601763409, | |
| "grad_norm": 1.3720532658754991, | |
| "learning_rate": 6.648984545254216e-07, | |
| "loss": 1.1034, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.4129316678912564, | |
| "grad_norm": 1.3638059704989733, | |
| "learning_rate": 6.626463415402131e-07, | |
| "loss": 1.1033, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.41440117560617196, | |
| "grad_norm": 1.3293227220141846, | |
| "learning_rate": 6.603905322109138e-07, | |
| "loss": 1.1054, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.41587068332108745, | |
| "grad_norm": 1.4610059929514447, | |
| "learning_rate": 6.581310778036474e-07, | |
| "loss": 0.9441, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.41734019103600295, | |
| "grad_norm": 1.412604693032899, | |
| "learning_rate": 6.558680296673766e-07, | |
| "loss": 1.1464, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.41880969875091845, | |
| "grad_norm": 1.3384277299187781, | |
| "learning_rate": 6.536014392327365e-07, | |
| "loss": 1.1396, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.42027920646583394, | |
| "grad_norm": 1.3987728346478638, | |
| "learning_rate": 6.51331358010865e-07, | |
| "loss": 1.0833, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.42174871418074944, | |
| "grad_norm": 1.577477591087323, | |
| "learning_rate": 6.490578375922328e-07, | |
| "loss": 1.1428, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.42321822189566494, | |
| "grad_norm": 1.5588639617673477, | |
| "learning_rate": 6.467809296454708e-07, | |
| "loss": 1.0857, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.42468772961058043, | |
| "grad_norm": 1.4118713026809537, | |
| "learning_rate": 6.445006859161956e-07, | |
| "loss": 1.1891, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.426157237325496, | |
| "grad_norm": 1.414060079207387, | |
| "learning_rate": 6.422171582258334e-07, | |
| "loss": 1.0867, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.4276267450404115, | |
| "grad_norm": 1.4123941837634646, | |
| "learning_rate": 6.399303984704432e-07, | |
| "loss": 1.1262, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.429096252755327, | |
| "grad_norm": 1.445359468963558, | |
| "learning_rate": 6.376404586195364e-07, | |
| "loss": 1.0714, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.4305657604702425, | |
| "grad_norm": 1.4230042553733626, | |
| "learning_rate": 6.353473907148961e-07, | |
| "loss": 1.1363, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.432035268185158, | |
| "grad_norm": 1.325942136523235, | |
| "learning_rate": 6.330512468693944e-07, | |
| "loss": 1.062, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.43350477590007347, | |
| "grad_norm": 1.4212600729175167, | |
| "learning_rate": 6.307520792658081e-07, | |
| "loss": 1.1282, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.43497428361498897, | |
| "grad_norm": 1.4254820055813413, | |
| "learning_rate": 6.284499401556328e-07, | |
| "loss": 1.1743, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.43644379132990446, | |
| "grad_norm": 1.4651887201716516, | |
| "learning_rate": 6.261448818578952e-07, | |
| "loss": 1.0821, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.43791329904481996, | |
| "grad_norm": 1.360035413094489, | |
| "learning_rate": 6.238369567579642e-07, | |
| "loss": 1.1277, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.4393828067597355, | |
| "grad_norm": 1.3842334621043852, | |
| "learning_rate": 6.215262173063607e-07, | |
| "loss": 1.134, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.440852314474651, | |
| "grad_norm": 1.5072722906394864, | |
| "learning_rate": 6.192127160175649e-07, | |
| "loss": 1.1385, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.4423218221895665, | |
| "grad_norm": 1.3613054784017538, | |
| "learning_rate": 6.168965054688238e-07, | |
| "loss": 1.1117, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.443791329904482, | |
| "grad_norm": 1.4231270036725268, | |
| "learning_rate": 6.145776382989552e-07, | |
| "loss": 1.0889, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.4452608376193975, | |
| "grad_norm": 1.4588262408030779, | |
| "learning_rate": 6.122561672071521e-07, | |
| "loss": 1.0359, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.446730345334313, | |
| "grad_norm": 1.411260640294453, | |
| "learning_rate": 6.099321449517851e-07, | |
| "loss": 1.1288, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.4481998530492285, | |
| "grad_norm": 1.3956027122525714, | |
| "learning_rate": 6.076056243492035e-07, | |
| "loss": 1.1173, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.449669360764144, | |
| "grad_norm": 1.4227235203095976, | |
| "learning_rate": 6.052766582725339e-07, | |
| "loss": 1.0394, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.45113886847905954, | |
| "grad_norm": 1.3911528073579336, | |
| "learning_rate": 6.029452996504801e-07, | |
| "loss": 1.1346, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.45260837619397504, | |
| "grad_norm": 1.3598137504924201, | |
| "learning_rate": 6.006116014661191e-07, | |
| "loss": 1.1715, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.45407788390889053, | |
| "grad_norm": 1.3341402291681796, | |
| "learning_rate": 5.982756167556978e-07, | |
| "loss": 1.0628, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.45554739162380603, | |
| "grad_norm": 1.4336307282930423, | |
| "learning_rate": 5.959373986074269e-07, | |
| "loss": 1.1031, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.4570168993387215, | |
| "grad_norm": 1.4435742655464976, | |
| "learning_rate": 5.935970001602751e-07, | |
| "loss": 1.0705, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.458486407053637, | |
| "grad_norm": 1.295586975475687, | |
| "learning_rate": 5.912544746027612e-07, | |
| "loss": 1.082, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.4599559147685525, | |
| "grad_norm": 1.419413405526378, | |
| "learning_rate": 5.88909875171745e-07, | |
| "loss": 1.1675, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.461425422483468, | |
| "grad_norm": 1.437607892446711, | |
| "learning_rate": 5.865632551512175e-07, | |
| "loss": 1.0499, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.4628949301983835, | |
| "grad_norm": 1.308082043000319, | |
| "learning_rate": 5.842146678710911e-07, | |
| "loss": 1.1021, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.46436443791329907, | |
| "grad_norm": 1.4411999480884474, | |
| "learning_rate": 5.818641667059856e-07, | |
| "loss": 1.115, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.46583394562821456, | |
| "grad_norm": 1.4248859054466834, | |
| "learning_rate": 5.795118050740169e-07, | |
| "loss": 1.0735, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.46730345334313006, | |
| "grad_norm": 1.3950677919239058, | |
| "learning_rate": 5.771576364355819e-07, | |
| "loss": 1.1289, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.46877296105804556, | |
| "grad_norm": 1.4356515357949984, | |
| "learning_rate": 5.748017142921448e-07, | |
| "loss": 1.1771, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.47024246877296105, | |
| "grad_norm": 1.3803272984145392, | |
| "learning_rate": 5.724440921850195e-07, | |
| "loss": 1.2042, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.47171197648787655, | |
| "grad_norm": 1.3828275996220702, | |
| "learning_rate": 5.700848236941543e-07, | |
| "loss": 1.1014, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.47318148420279205, | |
| "grad_norm": 1.4372467929841282, | |
| "learning_rate": 5.677239624369134e-07, | |
| "loss": 1.0635, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.47465099191770754, | |
| "grad_norm": 1.4412482068570687, | |
| "learning_rate": 5.653615620668589e-07, | |
| "loss": 1.0709, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.4761204996326231, | |
| "grad_norm": 1.5223869305783915, | |
| "learning_rate": 5.629976762725307e-07, | |
| "loss": 1.0613, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.4775900073475386, | |
| "grad_norm": 1.4106384412205724, | |
| "learning_rate": 5.606323587762275e-07, | |
| "loss": 1.3318, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.4790595150624541, | |
| "grad_norm": 1.4808028294260986, | |
| "learning_rate": 5.582656633327848e-07, | |
| "loss": 1.1348, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.4805290227773696, | |
| "grad_norm": 1.411232450754018, | |
| "learning_rate": 5.558976437283535e-07, | |
| "loss": 1.1455, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.4819985304922851, | |
| "grad_norm": 1.3874875642549753, | |
| "learning_rate": 5.535283537791785e-07, | |
| "loss": 1.1502, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.4834680382072006, | |
| "grad_norm": 1.529379416962907, | |
| "learning_rate": 5.511578473303742e-07, | |
| "loss": 1.1306, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.4849375459221161, | |
| "grad_norm": 1.4353633142872315, | |
| "learning_rate": 5.487861782547017e-07, | |
| "loss": 1.1453, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.4864070536370316, | |
| "grad_norm": 1.3293126841762646, | |
| "learning_rate": 5.464134004513442e-07, | |
| "loss": 1.1422, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.4878765613519471, | |
| "grad_norm": 1.3853962798430477, | |
| "learning_rate": 5.440395678446825e-07, | |
| "loss": 1.1431, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.4893460690668626, | |
| "grad_norm": 1.394633338720216, | |
| "learning_rate": 5.416647343830687e-07, | |
| "loss": 1.0997, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.4908155767817781, | |
| "grad_norm": 1.3557987209086912, | |
| "learning_rate": 5.392889540376006e-07, | |
| "loss": 1.0817, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.4922850844966936, | |
| "grad_norm": 2.8096656165818645, | |
| "learning_rate": 5.369122808008955e-07, | |
| "loss": 1.101, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.4937545922116091, | |
| "grad_norm": 1.463042361823667, | |
| "learning_rate": 5.345347686858626e-07, | |
| "loss": 1.1002, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.4952240999265246, | |
| "grad_norm": 1.513004558088391, | |
| "learning_rate": 5.321564717244757e-07, | |
| "loss": 1.0977, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.4966936076414401, | |
| "grad_norm": 1.513966316502475, | |
| "learning_rate": 5.297774439665449e-07, | |
| "loss": 1.1179, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.4981631153563556, | |
| "grad_norm": 1.3659383294811929, | |
| "learning_rate": 5.273977394784892e-07, | |
| "loss": 1.1009, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.4996326230712711, | |
| "grad_norm": 1.4406236511077748, | |
| "learning_rate": 5.250174123421068e-07, | |
| "loss": 1.2597, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.5011021307861866, | |
| "grad_norm": 1.4120228303295752, | |
| "learning_rate": 5.226365166533458e-07, | |
| "loss": 1.1148, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.5025716385011021, | |
| "grad_norm": 1.3150193319220376, | |
| "learning_rate": 5.202551065210768e-07, | |
| "loss": 1.0589, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.5040411462160176, | |
| "grad_norm": 1.3552885803220251, | |
| "learning_rate": 5.178732360658605e-07, | |
| "loss": 1.1281, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.5055106539309331, | |
| "grad_norm": 1.3782165714758237, | |
| "learning_rate": 5.154909594187192e-07, | |
| "loss": 1.115, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.5069801616458487, | |
| "grad_norm": 1.3839651456991835, | |
| "learning_rate": 5.131083307199071e-07, | |
| "loss": 1.1355, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.5084496693607642, | |
| "grad_norm": 1.3239848936899146, | |
| "learning_rate": 5.107254041176788e-07, | |
| "loss": 1.144, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.5099191770756797, | |
| "grad_norm": 1.4174744669080177, | |
| "learning_rate": 5.08342233767059e-07, | |
| "loss": 1.0891, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.5113886847905952, | |
| "grad_norm": 1.4197584070932192, | |
| "learning_rate": 5.059588738286118e-07, | |
| "loss": 1.1711, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.5128581925055107, | |
| "grad_norm": 1.460332635848889, | |
| "learning_rate": 5.035753784672105e-07, | |
| "loss": 1.1173, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.5143277002204262, | |
| "grad_norm": 1.4181458807955096, | |
| "learning_rate": 5.011918018508057e-07, | |
| "loss": 1.1409, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.5157972079353417, | |
| "grad_norm": 1.401868758466356, | |
| "learning_rate": 4.988081981491944e-07, | |
| "loss": 1.08, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.5172667156502572, | |
| "grad_norm": 1.4708608243218393, | |
| "learning_rate": 4.964246215327894e-07, | |
| "loss": 1.2088, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.5187362233651727, | |
| "grad_norm": 1.3087914829912093, | |
| "learning_rate": 4.940411261713882e-07, | |
| "loss": 1.088, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.5202057310800882, | |
| "grad_norm": 1.3639757292403556, | |
| "learning_rate": 4.91657766232941e-07, | |
| "loss": 1.1503, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.5216752387950037, | |
| "grad_norm": 1.3230263031119276, | |
| "learning_rate": 4.892745958823213e-07, | |
| "loss": 1.096, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.5231447465099192, | |
| "grad_norm": 1.3801068609972764, | |
| "learning_rate": 4.868916692800928e-07, | |
| "loss": 1.1064, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.5246142542248347, | |
| "grad_norm": 1.4818081892670187, | |
| "learning_rate": 4.845090405812809e-07, | |
| "loss": 1.0788, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.5260837619397501, | |
| "grad_norm": 1.4257930922965831, | |
| "learning_rate": 4.821267639341397e-07, | |
| "loss": 1.0475, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.5275532696546656, | |
| "grad_norm": 1.3664827162861548, | |
| "learning_rate": 4.797448934789232e-07, | |
| "loss": 1.0764, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.5290227773695811, | |
| "grad_norm": 1.384036889671994, | |
| "learning_rate": 4.773634833466541e-07, | |
| "loss": 1.0491, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.5304922850844966, | |
| "grad_norm": 1.3641782448500568, | |
| "learning_rate": 4.7498258765789335e-07, | |
| "loss": 1.0597, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.5319617927994122, | |
| "grad_norm": 1.423492847149906, | |
| "learning_rate": 4.726022605215108e-07, | |
| "loss": 1.2022, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.5334313005143277, | |
| "grad_norm": 1.4584247852241217, | |
| "learning_rate": 4.7022255603345504e-07, | |
| "loss": 1.0439, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.5349008082292432, | |
| "grad_norm": 1.4065237109476447, | |
| "learning_rate": 4.6784352827552433e-07, | |
| "loss": 1.0945, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.5363703159441587, | |
| "grad_norm": 1.359008408282674, | |
| "learning_rate": 4.6546523131413737e-07, | |
| "loss": 1.0858, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.5378398236590742, | |
| "grad_norm": 1.395920829743356, | |
| "learning_rate": 4.6308771919910455e-07, | |
| "loss": 1.0821, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.5393093313739897, | |
| "grad_norm": 1.4101595258715152, | |
| "learning_rate": 4.607110459623994e-07, | |
| "loss": 1.0566, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.5407788390889052, | |
| "grad_norm": 1.4400551979551548, | |
| "learning_rate": 4.5833526561693146e-07, | |
| "loss": 1.0875, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.5422483468038207, | |
| "grad_norm": 1.4569650644657774, | |
| "learning_rate": 4.559604321553176e-07, | |
| "loss": 1.0945, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.5437178545187362, | |
| "grad_norm": 1.4073481731963642, | |
| "learning_rate": 4.535865995486559e-07, | |
| "loss": 1.1326, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.5451873622336517, | |
| "grad_norm": 1.3843052803402691, | |
| "learning_rate": 4.512138217452984e-07, | |
| "loss": 1.0445, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.5466568699485672, | |
| "grad_norm": 1.482116494998853, | |
| "learning_rate": 4.488421526696259e-07, | |
| "loss": 1.0871, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.5481263776634827, | |
| "grad_norm": 1.3248580273316146, | |
| "learning_rate": 4.464716462208216e-07, | |
| "loss": 1.0515, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.5495958853783982, | |
| "grad_norm": 1.412909824990006, | |
| "learning_rate": 4.441023562716464e-07, | |
| "loss": 1.1317, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.5510653930933137, | |
| "grad_norm": 1.3923300607754752, | |
| "learning_rate": 4.417343366672154e-07, | |
| "loss": 1.0727, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.5525349008082292, | |
| "grad_norm": 1.3989792278937028, | |
| "learning_rate": 4.393676412237726e-07, | |
| "loss": 1.1355, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.5540044085231447, | |
| "grad_norm": 1.3946392251166513, | |
| "learning_rate": 4.370023237274693e-07, | |
| "loss": 1.0268, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.5554739162380602, | |
| "grad_norm": 1.41416247304468, | |
| "learning_rate": 4.3463843793314123e-07, | |
| "loss": 1.0584, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.5569434239529758, | |
| "grad_norm": 1.413788607440302, | |
| "learning_rate": 4.322760375630867e-07, | |
| "loss": 1.1846, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.5584129316678913, | |
| "grad_norm": 1.3360695239659472, | |
| "learning_rate": 4.299151763058457e-07, | |
| "loss": 1.1525, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.5598824393828068, | |
| "grad_norm": 1.3928904548700607, | |
| "learning_rate": 4.2755590781498056e-07, | |
| "loss": 1.0295, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.5613519470977223, | |
| "grad_norm": 1.4059368688788658, | |
| "learning_rate": 4.251982857078553e-07, | |
| "loss": 1.0474, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.5628214548126378, | |
| "grad_norm": 1.5189950174163334, | |
| "learning_rate": 4.2284236356441817e-07, | |
| "loss": 1.1166, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.5642909625275533, | |
| "grad_norm": 1.5773519459794447, | |
| "learning_rate": 4.204881949259832e-07, | |
| "loss": 1.216, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.5657604702424688, | |
| "grad_norm": 1.36093051783186, | |
| "learning_rate": 4.181358332940144e-07, | |
| "loss": 1.1028, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.5672299779573843, | |
| "grad_norm": 1.371801222763194, | |
| "learning_rate": 4.157853321289089e-07, | |
| "loss": 1.1299, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.5686994856722998, | |
| "grad_norm": 1.4309586056121164, | |
| "learning_rate": 4.1343674484878236e-07, | |
| "loss": 1.0961, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.5701689933872153, | |
| "grad_norm": 1.3916946581373462, | |
| "learning_rate": 4.11090124828255e-07, | |
| "loss": 1.084, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.5716385011021308, | |
| "grad_norm": 1.3379667194162195, | |
| "learning_rate": 4.0874552539723873e-07, | |
| "loss": 1.0954, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.5731080088170463, | |
| "grad_norm": 1.4682746859969864, | |
| "learning_rate": 4.064029998397247e-07, | |
| "loss": 1.123, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.5745775165319618, | |
| "grad_norm": 1.3797386330911343, | |
| "learning_rate": 4.04062601392573e-07, | |
| "loss": 1.1146, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.5760470242468773, | |
| "grad_norm": 1.4080533490533538, | |
| "learning_rate": 4.017243832443021e-07, | |
| "loss": 1.1117, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.5775165319617928, | |
| "grad_norm": 1.4279995751396037, | |
| "learning_rate": 3.993883985338808e-07, | |
| "loss": 1.0809, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.5789860396767083, | |
| "grad_norm": 1.3175097784254468, | |
| "learning_rate": 3.9705470034951986e-07, | |
| "loss": 1.0567, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.5804555473916239, | |
| "grad_norm": 1.3671088379708558, | |
| "learning_rate": 3.9472334172746596e-07, | |
| "loss": 1.1201, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.5819250551065394, | |
| "grad_norm": 1.4728248748732817, | |
| "learning_rate": 3.9239437565079645e-07, | |
| "loss": 1.1274, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.5833945628214549, | |
| "grad_norm": 1.4007045355088419, | |
| "learning_rate": 3.900678550482147e-07, | |
| "loss": 1.1204, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.5848640705363704, | |
| "grad_norm": 1.3734381784823617, | |
| "learning_rate": 3.877438327928478e-07, | |
| "loss": 1.1203, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.5863335782512858, | |
| "grad_norm": 1.4011368340836778, | |
| "learning_rate": 3.854223617010448e-07, | |
| "loss": 1.0533, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.5878030859662013, | |
| "grad_norm": 1.419012334062334, | |
| "learning_rate": 3.8310349453117617e-07, | |
| "loss": 1.0898, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.5892725936811168, | |
| "grad_norm": 1.3485874322249378, | |
| "learning_rate": 3.8078728398243503e-07, | |
| "loss": 1.1228, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.5907421013960323, | |
| "grad_norm": 1.4516071343212644, | |
| "learning_rate": 3.784737826936393e-07, | |
| "loss": 1.0773, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.5922116091109478, | |
| "grad_norm": 1.2870748285607192, | |
| "learning_rate": 3.761630432420358e-07, | |
| "loss": 1.0489, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.5936811168258633, | |
| "grad_norm": 1.4253057600978802, | |
| "learning_rate": 3.7385511814210493e-07, | |
| "loss": 1.1059, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.5951506245407788, | |
| "grad_norm": 1.4362421490175297, | |
| "learning_rate": 3.715500598443672e-07, | |
| "loss": 1.0798, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.5966201322556943, | |
| "grad_norm": 1.3142558631846668, | |
| "learning_rate": 3.6924792073419193e-07, | |
| "loss": 1.095, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.5980896399706098, | |
| "grad_norm": 1.4180620072245396, | |
| "learning_rate": 3.6694875313060567e-07, | |
| "loss": 1.1618, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.5995591476855253, | |
| "grad_norm": 1.4440333294474847, | |
| "learning_rate": 3.646526092851039e-07, | |
| "loss": 1.1742, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.6010286554004408, | |
| "grad_norm": 1.365950277580121, | |
| "learning_rate": 3.623595413804636e-07, | |
| "loss": 1.1049, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.6024981631153563, | |
| "grad_norm": 1.3800394626203474, | |
| "learning_rate": 3.600696015295568e-07, | |
| "loss": 1.0812, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.6039676708302718, | |
| "grad_norm": 1.4309055297758246, | |
| "learning_rate": 3.577828417741665e-07, | |
| "loss": 1.1591, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.6054371785451874, | |
| "grad_norm": 1.4376053158203683, | |
| "learning_rate": 3.5549931408380446e-07, | |
| "loss": 1.1083, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.6069066862601029, | |
| "grad_norm": 1.378409944475253, | |
| "learning_rate": 3.5321907035452913e-07, | |
| "loss": 1.0625, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.6083761939750184, | |
| "grad_norm": 1.4032013297402026, | |
| "learning_rate": 3.509421624077672e-07, | |
| "loss": 1.099, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.6098457016899339, | |
| "grad_norm": 1.403659794537397, | |
| "learning_rate": 3.486686419891349e-07, | |
| "loss": 1.0488, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.6113152094048494, | |
| "grad_norm": 1.3788151443608003, | |
| "learning_rate": 3.4639856076726346e-07, | |
| "loss": 1.113, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.6127847171197649, | |
| "grad_norm": 1.4287954215942635, | |
| "learning_rate": 3.4413197033262343e-07, | |
| "loss": 1.0992, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.6142542248346804, | |
| "grad_norm": 1.439755106322846, | |
| "learning_rate": 3.4186892219635254e-07, | |
| "loss": 1.0662, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.6157237325495959, | |
| "grad_norm": 1.4201132330693378, | |
| "learning_rate": 3.396094677890862e-07, | |
| "loss": 1.1108, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.6171932402645114, | |
| "grad_norm": 1.4151859235424602, | |
| "learning_rate": 3.373536584597869e-07, | |
| "loss": 1.043, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.6186627479794269, | |
| "grad_norm": 1.3436922877791724, | |
| "learning_rate": 3.3510154547457845e-07, | |
| "loss": 1.1002, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.6201322556943424, | |
| "grad_norm": 1.3768814743005016, | |
| "learning_rate": 3.3285318001558076e-07, | |
| "loss": 1.099, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.6216017634092579, | |
| "grad_norm": 1.9818793407102133, | |
| "learning_rate": 3.306086131797462e-07, | |
| "loss": 1.101, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.6230712711241734, | |
| "grad_norm": 1.5501432398127912, | |
| "learning_rate": 3.283678959776986e-07, | |
| "loss": 1.1401, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.6245407788390889, | |
| "grad_norm": 1.3451070295410414, | |
| "learning_rate": 3.261310793325742e-07, | |
| "loss": 1.0917, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.6260102865540044, | |
| "grad_norm": 1.3998218088329646, | |
| "learning_rate": 3.23898214078864e-07, | |
| "loss": 1.1472, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.6274797942689199, | |
| "grad_norm": 1.4699494469190635, | |
| "learning_rate": 3.216693509612587e-07, | |
| "loss": 1.0937, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.6289493019838354, | |
| "grad_norm": 1.463414683070946, | |
| "learning_rate": 3.19444540633495e-07, | |
| "loss": 1.1389, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.630418809698751, | |
| "grad_norm": 1.4765841861514846, | |
| "learning_rate": 3.172238336572056e-07, | |
| "loss": 1.2141, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.6318883174136665, | |
| "grad_norm": 1.4207175135849464, | |
| "learning_rate": 3.1500728050076873e-07, | |
| "loss": 1.0713, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.633357825128582, | |
| "grad_norm": 1.3488069296637555, | |
| "learning_rate": 3.1279493153816183e-07, | |
| "loss": 1.0483, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.6348273328434975, | |
| "grad_norm": 1.4026690915785558, | |
| "learning_rate": 3.1058683704781707e-07, | |
| "loss": 1.0378, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.636296840558413, | |
| "grad_norm": 1.368530405030037, | |
| "learning_rate": 3.0838304721147803e-07, | |
| "loss": 1.0595, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.6377663482733285, | |
| "grad_norm": 1.3798655085044127, | |
| "learning_rate": 3.0618361211305956e-07, | |
| "loss": 1.1855, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.639235855988244, | |
| "grad_norm": 1.4358356459308954, | |
| "learning_rate": 3.0398858173750994e-07, | |
| "loss": 1.1314, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.6407053637031594, | |
| "grad_norm": 1.3911243386477645, | |
| "learning_rate": 3.0179800596967414e-07, | |
| "loss": 1.1479, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.6421748714180749, | |
| "grad_norm": 1.3787211953067466, | |
| "learning_rate": 2.996119345931607e-07, | |
| "loss": 1.0852, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.6436443791329904, | |
| "grad_norm": 1.3784361879287548, | |
| "learning_rate": 2.9743041728921004e-07, | |
| "loss": 1.0617, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.6451138868479059, | |
| "grad_norm": 1.4563498056244404, | |
| "learning_rate": 2.952535036355659e-07, | |
| "loss": 1.0172, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.6465833945628214, | |
| "grad_norm": 1.3911833970178598, | |
| "learning_rate": 2.930812431053477e-07, | |
| "loss": 1.1772, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.6480529022777369, | |
| "grad_norm": 1.456141852044334, | |
| "learning_rate": 2.9091368506592704e-07, | |
| "loss": 1.1163, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.6495224099926524, | |
| "grad_norm": 1.4409576849555352, | |
| "learning_rate": 2.8875087877780547e-07, | |
| "loss": 1.0099, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.6509919177075679, | |
| "grad_norm": 1.3739666211555663, | |
| "learning_rate": 2.865928733934951e-07, | |
| "loss": 1.074, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.6524614254224834, | |
| "grad_norm": 1.4333872958890426, | |
| "learning_rate": 2.844397179564009e-07, | |
| "loss": 1.119, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.6539309331373989, | |
| "grad_norm": 1.307442428138931, | |
| "learning_rate": 2.8229146139970725e-07, | |
| "loss": 1.0908, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.6554004408523145, | |
| "grad_norm": 1.364990935232205, | |
| "learning_rate": 2.8014815254526475e-07, | |
| "loss": 1.0506, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.65686994856723, | |
| "grad_norm": 1.4165052577271036, | |
| "learning_rate": 2.780098401024816e-07, | |
| "loss": 1.1602, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.6583394562821455, | |
| "grad_norm": 1.4602397372513622, | |
| "learning_rate": 2.7587657266721633e-07, | |
| "loss": 1.0126, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.659808963997061, | |
| "grad_norm": 1.3583523079909423, | |
| "learning_rate": 2.737483987206725e-07, | |
| "loss": 1.0467, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.6612784717119765, | |
| "grad_norm": 1.376858530036489, | |
| "learning_rate": 2.7162536662829836e-07, | |
| "loss": 1.0401, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.662747979426892, | |
| "grad_norm": 1.403999668213476, | |
| "learning_rate": 2.695075246386874e-07, | |
| "loss": 1.0438, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.6642174871418075, | |
| "grad_norm": 1.3971075292397948, | |
| "learning_rate": 2.673949208824804e-07, | |
| "loss": 1.0229, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.665686994856723, | |
| "grad_norm": 1.4601504097247877, | |
| "learning_rate": 2.6528760337127344e-07, | |
| "loss": 1.0741, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.6671565025716385, | |
| "grad_norm": 1.3882632206062708, | |
| "learning_rate": 2.6318561999652543e-07, | |
| "loss": 1.1148, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.668626010286554, | |
| "grad_norm": 1.4454566469056964, | |
| "learning_rate": 2.610890185284707e-07, | |
| "loss": 1.155, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.6700955180014695, | |
| "grad_norm": 1.4144159704417698, | |
| "learning_rate": 2.5899784661503306e-07, | |
| "loss": 1.1574, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.671565025716385, | |
| "grad_norm": 1.3227253044283795, | |
| "learning_rate": 2.569121517807421e-07, | |
| "loss": 1.0814, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.6730345334313005, | |
| "grad_norm": 1.4016559184728539, | |
| "learning_rate": 2.5483198142565454e-07, | |
| "loss": 1.1234, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.674504041146216, | |
| "grad_norm": 1.407280713108974, | |
| "learning_rate": 2.5275738282427627e-07, | |
| "loss": 1.1664, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.6759735488611315, | |
| "grad_norm": 1.4047211167155722, | |
| "learning_rate": 2.506884031244875e-07, | |
| "loss": 1.0663, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.677443056576047, | |
| "grad_norm": 1.4210107328567902, | |
| "learning_rate": 2.4862508934647215e-07, | |
| "loss": 1.0806, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.6789125642909625, | |
| "grad_norm": 1.3638119047997457, | |
| "learning_rate": 2.465674883816492e-07, | |
| "loss": 1.1283, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.6803820720058781, | |
| "grad_norm": 1.3846202967932413, | |
| "learning_rate": 2.445156469916059e-07, | |
| "loss": 1.0391, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.6818515797207936, | |
| "grad_norm": 1.4492638193493528, | |
| "learning_rate": 2.424696118070367e-07, | |
| "loss": 1.1258, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.6833210874357091, | |
| "grad_norm": 1.3695917043777497, | |
| "learning_rate": 2.404294293266823e-07, | |
| "loss": 1.0027, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.6847905951506246, | |
| "grad_norm": 1.3993284836421556, | |
| "learning_rate": 2.3839514591627298e-07, | |
| "loss": 1.1171, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.6862601028655401, | |
| "grad_norm": 1.3323985034089447, | |
| "learning_rate": 2.3636680780747574e-07, | |
| "loss": 1.1226, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.6877296105804556, | |
| "grad_norm": 1.3810241812868833, | |
| "learning_rate": 2.3434446109684303e-07, | |
| "loss": 1.1231, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.6891991182953711, | |
| "grad_norm": 1.3868467700476377, | |
| "learning_rate": 2.323281517447646e-07, | |
| "loss": 1.1609, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.6906686260102866, | |
| "grad_norm": 1.4013984837064835, | |
| "learning_rate": 2.3031792557442426e-07, | |
| "loss": 1.0457, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.692138133725202, | |
| "grad_norm": 1.4712416790362257, | |
| "learning_rate": 2.2831382827075758e-07, | |
| "loss": 1.0643, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.6936076414401176, | |
| "grad_norm": 1.3781792405401019, | |
| "learning_rate": 2.2631590537941348e-07, | |
| "loss": 1.0947, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.695077149155033, | |
| "grad_norm": 1.4117456798087704, | |
| "learning_rate": 2.2432420230572014e-07, | |
| "loss": 1.1242, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.6965466568699485, | |
| "grad_norm": 1.3486912287876163, | |
| "learning_rate": 2.223387643136524e-07, | |
| "loss": 1.1288, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.698016164584864, | |
| "grad_norm": 1.4003361923857591, | |
| "learning_rate": 2.2035963652480266e-07, | |
| "loss": 1.0847, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.6994856722997795, | |
| "grad_norm": 1.4405126411409146, | |
| "learning_rate": 2.183868639173568e-07, | |
| "loss": 1.0318, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.700955180014695, | |
| "grad_norm": 1.3517140108244523, | |
| "learning_rate": 2.1642049132507013e-07, | |
| "loss": 1.1525, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.7024246877296105, | |
| "grad_norm": 1.401420474546432, | |
| "learning_rate": 2.144605634362504e-07, | |
| "loss": 1.0637, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.7038941954445261, | |
| "grad_norm": 1.4272891896192437, | |
| "learning_rate": 2.125071247927412e-07, | |
| "loss": 1.0649, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.7053637031594416, | |
| "grad_norm": 1.4147072802744984, | |
| "learning_rate": 2.1056021978890915e-07, | |
| "loss": 1.0819, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.7068332108743571, | |
| "grad_norm": 1.489071300786935, | |
| "learning_rate": 2.0861989267063622e-07, | |
| "loss": 1.1086, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.7083027185892726, | |
| "grad_norm": 1.3674178536485293, | |
| "learning_rate": 2.0668618753431372e-07, | |
| "loss": 1.126, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.7097722263041881, | |
| "grad_norm": 1.4610080335049196, | |
| "learning_rate": 2.0475914832583936e-07, | |
| "loss": 1.1212, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.7112417340191036, | |
| "grad_norm": 1.3760740804936542, | |
| "learning_rate": 2.0283881883961978e-07, | |
| "loss": 1.1276, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.7127112417340191, | |
| "grad_norm": 1.3670564421892966, | |
| "learning_rate": 2.0092524271757472e-07, | |
| "loss": 1.0993, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.7141807494489346, | |
| "grad_norm": 1.394984491530565, | |
| "learning_rate": 1.990184634481446e-07, | |
| "loss": 1.089, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.7156502571638501, | |
| "grad_norm": 1.4571577516475638, | |
| "learning_rate": 1.9711852436530318e-07, | |
| "loss": 1.125, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.7171197648787656, | |
| "grad_norm": 1.5499839330419916, | |
| "learning_rate": 1.952254686475726e-07, | |
| "loss": 1.1264, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.7185892725936811, | |
| "grad_norm": 1.3781629330474126, | |
| "learning_rate": 1.9333933931704098e-07, | |
| "loss": 1.0566, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.7200587803085966, | |
| "grad_norm": 1.4387738954738303, | |
| "learning_rate": 1.914601792383862e-07, | |
| "loss": 1.0665, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.7215282880235121, | |
| "grad_norm": 1.3566125913075402, | |
| "learning_rate": 1.8958803111790105e-07, | |
| "loss": 1.0703, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.7229977957384276, | |
| "grad_norm": 1.3556380519533342, | |
| "learning_rate": 1.877229375025222e-07, | |
| "loss": 1.1557, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.7244673034533431, | |
| "grad_norm": 1.3370842913006449, | |
| "learning_rate": 1.8586494077886416e-07, | |
| "loss": 1.098, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.7259368111682586, | |
| "grad_norm": 1.5097558318383415, | |
| "learning_rate": 1.840140831722557e-07, | |
| "loss": 1.1044, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.7274063188831741, | |
| "grad_norm": 1.4260075475587533, | |
| "learning_rate": 1.821704067457795e-07, | |
| "loss": 1.0891, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.7288758265980897, | |
| "grad_norm": 1.4163474844245585, | |
| "learning_rate": 1.803339533993175e-07, | |
| "loss": 1.1143, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.7303453343130052, | |
| "grad_norm": 1.3471773518411894, | |
| "learning_rate": 1.7850476486859784e-07, | |
| "loss": 1.1001, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.7318148420279207, | |
| "grad_norm": 1.349662778135853, | |
| "learning_rate": 1.766828827242461e-07, | |
| "loss": 1.0691, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.7332843497428362, | |
| "grad_norm": 1.413952156125001, | |
| "learning_rate": 1.7486834837084147e-07, | |
| "loss": 1.1791, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.7347538574577517, | |
| "grad_norm": 1.4517736063725735, | |
| "learning_rate": 1.7306120304597516e-07, | |
| "loss": 1.1076, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.7347538574577517, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_loss": 2.7672901153564453, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_runtime": 5.874, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_samples_per_second": 15.832, | |
| "eval_ical_mcts_chains_sft_val_MORECHAINS_steps_per_second": 2.043, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.7362233651726672, | |
| "grad_norm": 1.4017708695893076, | |
| "learning_rate": 1.7126148781931309e-07, | |
| "loss": 1.2015, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.7376928728875827, | |
| "grad_norm": 1.373359024365262, | |
| "learning_rate": 1.6946924359166332e-07, | |
| "loss": 1.1271, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.7391623806024982, | |
| "grad_norm": 1.3527598369707952, | |
| "learning_rate": 1.6768451109404518e-07, | |
| "loss": 1.1452, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.7406318883174137, | |
| "grad_norm": 1.4342813387133044, | |
| "learning_rate": 1.659073308867653e-07, | |
| "loss": 1.0873, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.7421013960323292, | |
| "grad_norm": 1.3934397068509279, | |
| "learning_rate": 1.641377433584945e-07, | |
| "loss": 1.1442, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.7435709037472447, | |
| "grad_norm": 1.3468387357646852, | |
| "learning_rate": 1.6237578872535023e-07, | |
| "loss": 1.0956, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.7450404114621602, | |
| "grad_norm": 1.3379393102784192, | |
| "learning_rate": 1.6062150702998307e-07, | |
| "loss": 1.1059, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.7465099191770757, | |
| "grad_norm": 1.4208106676166712, | |
| "learning_rate": 1.5887493814066632e-07, | |
| "loss": 1.0594, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.7479794268919912, | |
| "grad_norm": 1.3284093386527018, | |
| "learning_rate": 1.5713612175038953e-07, | |
| "loss": 1.0709, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.7494489346069066, | |
| "grad_norm": 1.3401638532562588, | |
| "learning_rate": 1.5540509737595752e-07, | |
| "loss": 1.0478, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.7509184423218221, | |
| "grad_norm": 1.387779204411568, | |
| "learning_rate": 1.536819043570915e-07, | |
| "loss": 1.0749, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.7523879500367376, | |
| "grad_norm": 1.3361938042809636, | |
| "learning_rate": 1.5196658185553484e-07, | |
| "loss": 1.0633, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.7538574577516532, | |
| "grad_norm": 1.4464634414727173, | |
| "learning_rate": 1.5025916885416385e-07, | |
| "loss": 1.0915, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.7553269654665687, | |
| "grad_norm": 1.4564936678530274, | |
| "learning_rate": 1.485597041561014e-07, | |
| "loss": 1.181, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.7567964731814842, | |
| "grad_norm": 1.3813741239097441, | |
| "learning_rate": 1.4686822638383485e-07, | |
| "loss": 1.1353, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.7582659808963997, | |
| "grad_norm": 1.3284963146667195, | |
| "learning_rate": 1.4518477397833868e-07, | |
| "loss": 1.0876, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.7597354886113152, | |
| "grad_norm": 1.3825280093987735, | |
| "learning_rate": 1.4350938519820082e-07, | |
| "loss": 1.1206, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.7612049963262307, | |
| "grad_norm": 1.4124836120841386, | |
| "learning_rate": 1.4184209811875314e-07, | |
| "loss": 1.0521, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.7626745040411462, | |
| "grad_norm": 1.3957995246414707, | |
| "learning_rate": 1.401829506312061e-07, | |
| "loss": 1.1441, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.7641440117560617, | |
| "grad_norm": 2.064618166415374, | |
| "learning_rate": 1.385319804417872e-07, | |
| "loss": 1.0951, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.7656135194709772, | |
| "grad_norm": 1.3751431109514682, | |
| "learning_rate": 1.3688922507088506e-07, | |
| "loss": 1.0622, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.7670830271858927, | |
| "grad_norm": 1.4467861329866494, | |
| "learning_rate": 1.35254721852196e-07, | |
| "loss": 1.1695, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.7685525349008082, | |
| "grad_norm": 1.37927918872005, | |
| "learning_rate": 1.3362850793187536e-07, | |
| "loss": 1.1666, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.7700220426157237, | |
| "grad_norm": 1.402761123927447, | |
| "learning_rate": 1.3201062026769415e-07, | |
| "loss": 1.1472, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.7714915503306392, | |
| "grad_norm": 1.4020269134593764, | |
| "learning_rate": 1.3040109562819852e-07, | |
| "loss": 1.1132, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.7729610580455547, | |
| "grad_norm": 1.3179769606979768, | |
| "learning_rate": 1.2879997059187402e-07, | |
| "loss": 1.1577, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.7744305657604702, | |
| "grad_norm": 1.3608872827916707, | |
| "learning_rate": 1.27207281546315e-07, | |
| "loss": 1.105, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.7759000734753857, | |
| "grad_norm": 1.4191003452870128, | |
| "learning_rate": 1.2562306468739707e-07, | |
| "loss": 1.061, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.7773695811903012, | |
| "grad_norm": 1.42228631098522, | |
| "learning_rate": 1.2404735601845446e-07, | |
| "loss": 1.0901, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.7788390889052168, | |
| "grad_norm": 1.374432034447507, | |
| "learning_rate": 1.2248019134946224e-07, | |
| "loss": 1.1478, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.7803085966201323, | |
| "grad_norm": 1.4436649330903892, | |
| "learning_rate": 1.2092160629622243e-07, | |
| "loss": 1.2448, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.7817781043350478, | |
| "grad_norm": 1.3822744122903736, | |
| "learning_rate": 1.1937163627955388e-07, | |
| "loss": 1.0715, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.7832476120499633, | |
| "grad_norm": 1.4003007054983763, | |
| "learning_rate": 1.1783031652448844e-07, | |
| "loss": 1.0834, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.7847171197648788, | |
| "grad_norm": 1.3009002478832499, | |
| "learning_rate": 1.1629768205946916e-07, | |
| "loss": 1.0867, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.7861866274797943, | |
| "grad_norm": 1.384225049312056, | |
| "learning_rate": 1.1477376771555547e-07, | |
| "loss": 1.111, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.7876561351947098, | |
| "grad_norm": 1.4951881255718384, | |
| "learning_rate": 1.1325860812563082e-07, | |
| "loss": 1.0905, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.7891256429096253, | |
| "grad_norm": 1.3855016278493693, | |
| "learning_rate": 1.1175223772361548e-07, | |
| "loss": 1.1674, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.7905951506245408, | |
| "grad_norm": 1.402883620503283, | |
| "learning_rate": 1.1025469074368465e-07, | |
| "loss": 1.0934, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.7920646583394563, | |
| "grad_norm": 1.3721282887774249, | |
| "learning_rate": 1.0876600121949014e-07, | |
| "loss": 1.053, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.7935341660543718, | |
| "grad_norm": 1.3027250754946427, | |
| "learning_rate": 1.0728620298338647e-07, | |
| "loss": 0.9932, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.7950036737692873, | |
| "grad_norm": 1.3472091716626784, | |
| "learning_rate": 1.058153296656627e-07, | |
| "loss": 1.0411, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.7964731814842028, | |
| "grad_norm": 1.2515159561759799, | |
| "learning_rate": 1.0435341469377785e-07, | |
| "loss": 1.0409, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.7979426891991183, | |
| "grad_norm": 1.3955411540097424, | |
| "learning_rate": 1.0290049129160083e-07, | |
| "loss": 1.0847, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.7994121969140338, | |
| "grad_norm": 1.378388349777933, | |
| "learning_rate": 1.0145659247865606e-07, | |
| "loss": 1.0637, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.8008817046289493, | |
| "grad_norm": 1.3665203225260263, | |
| "learning_rate": 1.0002175106937282e-07, | |
| "loss": 1.083, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.8023512123438648, | |
| "grad_norm": 1.3325384707568575, | |
| "learning_rate": 9.859599967233901e-08, | |
| "loss": 1.0619, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.8038207200587804, | |
| "grad_norm": 1.3079374620261692, | |
| "learning_rate": 9.717937068956083e-08, | |
| "loss": 1.0893, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.8052902277736959, | |
| "grad_norm": 1.386818107212991, | |
| "learning_rate": 9.577189631572613e-08, | |
| "loss": 1.1115, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.8067597354886114, | |
| "grad_norm": 1.374316392770813, | |
| "learning_rate": 9.437360853747223e-08, | |
| "loss": 1.1449, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.8082292432035268, | |
| "grad_norm": 1.345287353676708, | |
| "learning_rate": 9.29845391326598e-08, | |
| "loss": 1.0576, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.8096987509184423, | |
| "grad_norm": 1.4813724241887494, | |
| "learning_rate": 9.16047196696505e-08, | |
| "loss": 1.092, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.8111682586333578, | |
| "grad_norm": 1.3806739470510292, | |
| "learning_rate": 9.023418150658863e-08, | |
| "loss": 1.0925, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.8126377663482733, | |
| "grad_norm": 1.3661938482257079, | |
| "learning_rate": 8.887295579068988e-08, | |
| "loss": 1.0278, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.8141072740631888, | |
| "grad_norm": 1.376286461998827, | |
| "learning_rate": 8.752107345753262e-08, | |
| "loss": 1.0203, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.8155767817781043, | |
| "grad_norm": 1.3714175983707415, | |
| "learning_rate": 8.617856523035466e-08, | |
| "loss": 1.0947, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.8170462894930198, | |
| "grad_norm": 1.3804861761749325, | |
| "learning_rate": 8.484546161935596e-08, | |
| "loss": 1.0772, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.8185157972079353, | |
| "grad_norm": 1.374313789743847, | |
| "learning_rate": 8.352179292100403e-08, | |
| "loss": 1.0653, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.8199853049228508, | |
| "grad_norm": 1.3618284844063149, | |
| "learning_rate": 8.220758921734649e-08, | |
| "loss": 1.0513, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.8214548126377663, | |
| "grad_norm": 1.3803405029205986, | |
| "learning_rate": 8.090288037532706e-08, | |
| "loss": 1.1005, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.8229243203526818, | |
| "grad_norm": 1.3988974278658783, | |
| "learning_rate": 7.960769604610618e-08, | |
| "loss": 1.1205, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.8243938280675973, | |
| "grad_norm": 1.3526087062032914, | |
| "learning_rate": 7.83220656643881e-08, | |
| "loss": 1.0322, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.8258633357825128, | |
| "grad_norm": 1.3213029228844033, | |
| "learning_rate": 7.704601844775155e-08, | |
| "loss": 1.071, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.8273328434974284, | |
| "grad_norm": 1.4988675658645414, | |
| "learning_rate": 7.577958339598529e-08, | |
| "loss": 1.0068, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.8288023512123439, | |
| "grad_norm": 1.4472214140414408, | |
| "learning_rate": 7.452278929042982e-08, | |
| "loss": 1.1286, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.8302718589272594, | |
| "grad_norm": 1.3652634375890536, | |
| "learning_rate": 7.327566469332303e-08, | |
| "loss": 1.1308, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.8317413666421749, | |
| "grad_norm": 1.3655886385805245, | |
| "learning_rate": 7.203823794715041e-08, | |
| "loss": 1.1015, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.8332108743570904, | |
| "grad_norm": 1.2502888959822873, | |
| "learning_rate": 7.08105371740021e-08, | |
| "loss": 1.0829, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.8346803820720059, | |
| "grad_norm": 1.366330743048358, | |
| "learning_rate": 6.959259027493303e-08, | |
| "loss": 1.1187, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.8361498897869214, | |
| "grad_norm": 1.3645999375102753, | |
| "learning_rate": 6.838442492932867e-08, | |
| "loss": 1.0662, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.8376193975018369, | |
| "grad_norm": 1.4139397317247722, | |
| "learning_rate": 6.718606859427673e-08, | |
| "loss": 1.1166, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.8390889052167524, | |
| "grad_norm": 1.4003308125547838, | |
| "learning_rate": 6.599754850394263e-08, | |
| "loss": 1.1006, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.8405584129316679, | |
| "grad_norm": 1.3719288368595044, | |
| "learning_rate": 6.481889166895033e-08, | |
| "loss": 1.0538, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.8420279206465834, | |
| "grad_norm": 1.3259310033546516, | |
| "learning_rate": 6.365012487576926e-08, | |
| "loss": 1.0573, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.8434974283614989, | |
| "grad_norm": 1.4183409718497981, | |
| "learning_rate": 6.249127468610504e-08, | |
| "loss": 1.1412, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.8449669360764144, | |
| "grad_norm": 1.3764865756818885, | |
| "learning_rate": 6.134236743629562e-08, | |
| "loss": 1.1218, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.8464364437913299, | |
| "grad_norm": 1.4471292335999226, | |
| "learning_rate": 6.020342923671334e-08, | |
| "loss": 1.0495, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.8479059515062454, | |
| "grad_norm": 1.4152415184940785, | |
| "learning_rate": 5.907448597117126e-08, | |
| "loss": 1.1413, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.8493754592211609, | |
| "grad_norm": 1.3749110659106119, | |
| "learning_rate": 5.7955563296334664e-08, | |
| "loss": 1.1659, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.8508449669360764, | |
| "grad_norm": 1.4403389990835915, | |
| "learning_rate": 5.6846686641138394e-08, | |
| "loss": 1.0724, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.852314474650992, | |
| "grad_norm": 1.3113544123287635, | |
| "learning_rate": 5.5747881206208936e-08, | |
| "loss": 1.1002, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.8537839823659075, | |
| "grad_norm": 1.4504739763414178, | |
| "learning_rate": 5.465917196329106e-08, | |
| "loss": 1.0829, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.855253490080823, | |
| "grad_norm": 1.461654698052296, | |
| "learning_rate": 5.3580583654681266e-08, | |
| "loss": 1.0559, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.8567229977957385, | |
| "grad_norm": 1.4101135770661162, | |
| "learning_rate": 5.251214079266475e-08, | |
| "loss": 1.069, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.858192505510654, | |
| "grad_norm": 1.3634335354582763, | |
| "learning_rate": 5.1453867658958704e-08, | |
| "loss": 1.1287, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.8596620132255695, | |
| "grad_norm": 1.4283589117852606, | |
| "learning_rate": 5.0405788304160426e-08, | |
| "loss": 1.0935, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.861131520940485, | |
| "grad_norm": 1.434406958645411, | |
| "learning_rate": 4.936792654720029e-08, | |
| "loss": 1.0494, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.8626010286554004, | |
| "grad_norm": 1.4389756558713183, | |
| "learning_rate": 4.8340305974801266e-08, | |
| "loss": 1.0656, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.864070536370316, | |
| "grad_norm": 1.3842315564330259, | |
| "learning_rate": 4.7322949940942325e-08, | |
| "loss": 1.0512, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.8655400440852314, | |
| "grad_norm": 1.4148624042139615, | |
| "learning_rate": 4.63158815663276e-08, | |
| "loss": 1.0613, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.8670095518001469, | |
| "grad_norm": 1.3366794867264846, | |
| "learning_rate": 4.53191237378614e-08, | |
| "loss": 1.0486, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.8684790595150624, | |
| "grad_norm": 1.3575430817405785, | |
| "learning_rate": 4.433269910812759e-08, | |
| "loss": 1.0735, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.8699485672299779, | |
| "grad_norm": 1.3289314595584882, | |
| "learning_rate": 4.335663009487511e-08, | |
| "loss": 1.1759, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.8714180749448934, | |
| "grad_norm": 1.4000173838833194, | |
| "learning_rate": 4.2390938880508595e-08, | |
| "loss": 1.0408, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.8728875826598089, | |
| "grad_norm": 1.372416148633152, | |
| "learning_rate": 4.143564741158362e-08, | |
| "loss": 1.0626, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.8743570903747244, | |
| "grad_norm": 1.4265940660232175, | |
| "learning_rate": 4.0490777398308753e-08, | |
| "loss": 1.2163, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.8758265980896399, | |
| "grad_norm": 1.3925507856636796, | |
| "learning_rate": 3.955635031405169e-08, | |
| "loss": 1.1028, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.8772961058045555, | |
| "grad_norm": 1.324576005377705, | |
| "learning_rate": 3.86323873948512e-08, | |
| "loss": 1.182, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.878765613519471, | |
| "grad_norm": 1.351029057416015, | |
| "learning_rate": 3.771890963893476e-08, | |
| "loss": 1.1285, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.8802351212343865, | |
| "grad_norm": 1.3994327447522252, | |
| "learning_rate": 3.681593780624137e-08, | |
| "loss": 1.0868, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.881704628949302, | |
| "grad_norm": 1.3561535752329212, | |
| "learning_rate": 3.5923492417949285e-08, | |
| "loss": 1.1401, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.8831741366642175, | |
| "grad_norm": 1.3988534133786137, | |
| "learning_rate": 3.5041593756010234e-08, | |
| "loss": 1.0106, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.884643644379133, | |
| "grad_norm": 1.3859821796457574, | |
| "learning_rate": 3.417026186268829e-08, | |
| "loss": 1.0873, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.8861131520940485, | |
| "grad_norm": 1.3465915197780411, | |
| "learning_rate": 3.3309516540104e-08, | |
| "loss": 1.1312, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.887582659808964, | |
| "grad_norm": 1.3934103882388418, | |
| "learning_rate": 3.2459377349784986e-08, | |
| "loss": 1.0802, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.8890521675238795, | |
| "grad_norm": 1.437718999786439, | |
| "learning_rate": 3.1619863612221075e-08, | |
| "loss": 1.182, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.890521675238795, | |
| "grad_norm": 1.4477961897782676, | |
| "learning_rate": 3.079099440642496e-08, | |
| "loss": 1.1751, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.8919911829537105, | |
| "grad_norm": 1.3924612008055068, | |
| "learning_rate": 2.997278856949914e-08, | |
| "loss": 1.0903, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.893460690668626, | |
| "grad_norm": 1.3907249036661788, | |
| "learning_rate": 2.916526469620756e-08, | |
| "loss": 1.0676, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.8949301983835415, | |
| "grad_norm": 1.4577046884258984, | |
| "learning_rate": 2.836844113855269e-08, | |
| "loss": 1.2377, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.896399706098457, | |
| "grad_norm": 1.4063512104136562, | |
| "learning_rate": 2.758233600535914e-08, | |
| "loss": 1.0655, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.8978692138133725, | |
| "grad_norm": 1.3811245692933716, | |
| "learning_rate": 2.6806967161861593e-08, | |
| "loss": 1.0515, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.899338721528288, | |
| "grad_norm": 1.4687882128856604, | |
| "learning_rate": 2.6042352229298902e-08, | |
| "loss": 1.0785, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.9008082292432035, | |
| "grad_norm": 1.4655898408700718, | |
| "learning_rate": 2.5288508584513814e-08, | |
| "loss": 1.0829, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.9022777369581191, | |
| "grad_norm": 1.3913574357745526, | |
| "learning_rate": 2.4545453359557765e-08, | |
| "loss": 1.0768, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.9037472446730346, | |
| "grad_norm": 1.3607919197214273, | |
| "learning_rate": 2.3813203441301778e-08, | |
| "loss": 1.1146, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.9052167523879501, | |
| "grad_norm": 1.4196701845145032, | |
| "learning_rate": 2.3091775471052734e-08, | |
| "loss": 1.1263, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.9066862601028656, | |
| "grad_norm": 1.3646134363098568, | |
| "learning_rate": 2.2381185844174644e-08, | |
| "loss": 1.027, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.9081557678177811, | |
| "grad_norm": 1.4522612796957168, | |
| "learning_rate": 2.168145070971683e-08, | |
| "loss": 1.0925, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.9096252755326966, | |
| "grad_norm": 1.332216974123385, | |
| "learning_rate": 2.099258597004644e-08, | |
| "loss": 1.057, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.9110947832476121, | |
| "grad_norm": 1.4818724681035569, | |
| "learning_rate": 2.031460728048695e-08, | |
| "loss": 0.9901, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.9125642909625276, | |
| "grad_norm": 1.3874842041637756, | |
| "learning_rate": 1.9647530048962747e-08, | |
| "loss": 1.0946, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.914033798677443, | |
| "grad_norm": 1.3490442752252867, | |
| "learning_rate": 1.8991369435648774e-08, | |
| "loss": 1.1295, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.9155033063923586, | |
| "grad_norm": 1.3339563383295754, | |
| "learning_rate": 1.8346140352625883e-08, | |
| "loss": 1.0182, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.916972814107274, | |
| "grad_norm": 1.4464185646745273, | |
| "learning_rate": 1.771185746354209e-08, | |
| "loss": 1.0926, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.9184423218221895, | |
| "grad_norm": 1.3165206597074977, | |
| "learning_rate": 1.7088535183279407e-08, | |
| "loss": 1.1249, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.919911829537105, | |
| "grad_norm": 1.3819515848899369, | |
| "learning_rate": 1.647618767762593e-08, | |
| "loss": 1.1573, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.9213813372520205, | |
| "grad_norm": 1.3001816716305787, | |
| "learning_rate": 1.5874828862954327e-08, | |
| "loss": 1.0266, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.922850844966936, | |
| "grad_norm": 1.319276221215603, | |
| "learning_rate": 1.5284472405905247e-08, | |
| "loss": 1.0694, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.9243203526818515, | |
| "grad_norm": 1.4228373976898447, | |
| "learning_rate": 1.4705131723076692e-08, | |
| "loss": 1.1516, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.925789860396767, | |
| "grad_norm": 1.3595993483558204, | |
| "learning_rate": 1.4136819980719472e-08, | |
| "loss": 1.0336, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.9272593681116826, | |
| "grad_norm": 1.3326370942028576, | |
| "learning_rate": 1.3579550094437676e-08, | |
| "loss": 1.0843, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.9287288758265981, | |
| "grad_norm": 1.3933226824191673, | |
| "learning_rate": 1.3033334728895119e-08, | |
| "loss": 1.1692, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.9301983835415136, | |
| "grad_norm": 1.3772541537537315, | |
| "learning_rate": 1.2498186297527802e-08, | |
| "loss": 1.1253, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.9316678912564291, | |
| "grad_norm": 1.4025797539924674, | |
| "learning_rate": 1.1974116962261527e-08, | |
| "loss": 1.1508, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.9331373989713446, | |
| "grad_norm": 1.3051053877123164, | |
| "learning_rate": 1.1461138633235611e-08, | |
| "loss": 1.0791, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.9346069066862601, | |
| "grad_norm": 1.3333191944243534, | |
| "learning_rate": 1.095926296853228e-08, | |
| "loss": 1.1968, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.9360764144011756, | |
| "grad_norm": 1.3418559426331282, | |
| "learning_rate": 1.0468501373911532e-08, | |
| "loss": 1.0896, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.9375459221160911, | |
| "grad_norm": 1.34359709316688, | |
| "learning_rate": 9.988865002552138e-09, | |
| "loss": 1.1065, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.9390154298310066, | |
| "grad_norm": 1.3620227287509608, | |
| "learning_rate": 9.520364754798116e-09, | |
| "loss": 1.0744, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.9404849375459221, | |
| "grad_norm": 1.3087398924414495, | |
| "learning_rate": 9.06301127791087e-09, | |
| "loss": 1.072, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.9419544452608376, | |
| "grad_norm": 1.4212527547101943, | |
| "learning_rate": 8.61681496582739e-09, | |
| "loss": 1.0904, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.9434239529757531, | |
| "grad_norm": 1.4044007797229823, | |
| "learning_rate": 8.181785958923938e-09, | |
| "loss": 1.1127, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.9448934606906686, | |
| "grad_norm": 1.3492187382144503, | |
| "learning_rate": 7.757934143785561e-09, | |
| "loss": 1.1309, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.9463629684055841, | |
| "grad_norm": 1.3882974174543632, | |
| "learning_rate": 7.345269152981614e-09, | |
| "loss": 1.0746, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.9478324761204996, | |
| "grad_norm": 1.3919108675813483, | |
| "learning_rate": 6.943800364846653e-09, | |
| "loss": 1.1149, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.9493019838354151, | |
| "grad_norm": 1.4427576763993042, | |
| "learning_rate": 6.5535369032672095e-09, | |
| "loss": 1.1289, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.9507714915503307, | |
| "grad_norm": 1.4253308850896702, | |
| "learning_rate": 6.174487637474801e-09, | |
| "loss": 1.1042, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.9522409992652462, | |
| "grad_norm": 1.400034959006393, | |
| "learning_rate": 5.806661181843919e-09, | |
| "loss": 1.0803, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.9537105069801617, | |
| "grad_norm": 1.435130822553809, | |
| "learning_rate": 5.450065895696632e-09, | |
| "loss": 1.0796, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.9551800146950772, | |
| "grad_norm": 1.4035193875819156, | |
| "learning_rate": 5.1047098831125124e-09, | |
| "loss": 1.1129, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.9566495224099927, | |
| "grad_norm": 1.5238894343006477, | |
| "learning_rate": 4.770600992744178e-09, | |
| "loss": 1.1102, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.9581190301249082, | |
| "grad_norm": 1.3438191535373563, | |
| "learning_rate": 4.4477468176393196e-09, | |
| "loss": 1.0581, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.9595885378398237, | |
| "grad_norm": 1.3010475336208236, | |
| "learning_rate": 4.136154695068006e-09, | |
| "loss": 1.0478, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.9610580455547392, | |
| "grad_norm": 1.3915831877261908, | |
| "learning_rate": 3.8358317063557635e-09, | |
| "loss": 1.0616, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.9625275532696547, | |
| "grad_norm": 1.3754622477764384, | |
| "learning_rate": 3.546784676722925e-09, | |
| "loss": 1.0684, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.9639970609845702, | |
| "grad_norm": 1.3958030478092502, | |
| "learning_rate": 3.2690201751292002e-09, | |
| "loss": 1.1125, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.9654665686994857, | |
| "grad_norm": 1.4118012322738474, | |
| "learning_rate": 3.002544514124683e-09, | |
| "loss": 1.0673, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.9669360764144012, | |
| "grad_norm": 1.3931161551594535, | |
| "learning_rate": 2.747363749706244e-09, | |
| "loss": 1.1434, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.9684055841293167, | |
| "grad_norm": 1.320233907465206, | |
| "learning_rate": 2.5034836811799744e-09, | |
| "loss": 1.0626, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.9698750918442322, | |
| "grad_norm": 1.40459539845844, | |
| "learning_rate": 2.2709098510292347e-09, | |
| "loss": 1.1047, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.9713445995591476, | |
| "grad_norm": 1.4426402012148907, | |
| "learning_rate": 2.049647544788813e-09, | |
| "loss": 1.0249, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.9728141072740631, | |
| "grad_norm": 1.37434918164306, | |
| "learning_rate": 1.8397017909249634e-09, | |
| "loss": 1.0911, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.9742836149889786, | |
| "grad_norm": 1.4108384424402636, | |
| "learning_rate": 1.6410773607206663e-09, | |
| "loss": 1.1213, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.9757531227038942, | |
| "grad_norm": 1.4926584455705627, | |
| "learning_rate": 1.4537787681677683e-09, | |
| "loss": 1.1082, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.9772226304188097, | |
| "grad_norm": 1.3595606693253082, | |
| "learning_rate": 1.2778102698638993e-09, | |
| "loss": 1.0751, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.9786921381337252, | |
| "grad_norm": 1.4509988964421765, | |
| "learning_rate": 1.1131758649160494e-09, | |
| "loss": 1.0865, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.9801616458486407, | |
| "grad_norm": 1.3679120087919863, | |
| "learning_rate": 9.598792948496414e-10, | |
| "loss": 1.0855, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.9816311535635562, | |
| "grad_norm": 1.346685189196511, | |
| "learning_rate": 8.179240435232659e-10, | |
| "loss": 1.0846, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.9831006612784717, | |
| "grad_norm": 1.3196217753280912, | |
| "learning_rate": 6.873133370498551e-10, | |
| "loss": 1.0291, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.9845701689933872, | |
| "grad_norm": 1.3795048616719339, | |
| "learning_rate": 5.680501437230755e-10, | |
| "loss": 1.0951, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.9860396767083027, | |
| "grad_norm": 1.3485249223159093, | |
| "learning_rate": 4.6013717395010365e-10, | |
| "loss": 1.0971, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.9875091844232182, | |
| "grad_norm": 1.507892762769766, | |
| "learning_rate": 3.63576880189731e-10, | |
| "loss": 1.1302, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.9889786921381337, | |
| "grad_norm": 1.362116408246938, | |
| "learning_rate": 2.783714568970197e-10, | |
| "loss": 1.1335, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.9904481998530492, | |
| "grad_norm": 1.4376872456307725, | |
| "learning_rate": 2.045228404731203e-10, | |
| "loss": 1.0601, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.9919177075679647, | |
| "grad_norm": 1.3558644824462847, | |
| "learning_rate": 1.4203270922125143e-10, | |
| "loss": 1.0517, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.9933872152828802, | |
| "grad_norm": 1.4975627645392091, | |
| "learning_rate": 9.090248330889671e-11, | |
| "loss": 1.0872, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.9948567229977957, | |
| "grad_norm": 1.5061190709997394, | |
| "learning_rate": 5.1133324735164183e-11, | |
| "loss": 1.0224, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.9963262307127112, | |
| "grad_norm": 1.4530193394151236, | |
| "learning_rate": 2.2726137304529546e-11, | |
| "loss": 1.0538, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.9977957384276267, | |
| "grad_norm": 1.3983394789120556, | |
| "learning_rate": 5.6815666063525505e-12, | |
| "loss": 1.0876, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.9992652461425422, | |
| "grad_norm": 1.4207479289834348, | |
| "learning_rate": 0.0, | |
| "loss": 1.1348, | |
| "step": 680 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 680, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 711821505200128.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |