| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 704, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002840909090909091, | |
| "grad_norm": 5.7846018213668176, | |
| "learning_rate": 0.0, | |
| "loss": 0.9187, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.005681818181818182, | |
| "grad_norm": 5.6013787706769165, | |
| "learning_rate": 5.681818181818182e-08, | |
| "loss": 0.9337, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.008522727272727272, | |
| "grad_norm": 5.628576098877568, | |
| "learning_rate": 1.1363636363636364e-07, | |
| "loss": 0.9727, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.011363636363636364, | |
| "grad_norm": 5.6880005354310885, | |
| "learning_rate": 1.7045454545454545e-07, | |
| "loss": 0.9255, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.014204545454545454, | |
| "grad_norm": 5.1670651022067124, | |
| "learning_rate": 2.2727272727272729e-07, | |
| "loss": 0.9236, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.017045454545454544, | |
| "grad_norm": 5.568976898855187, | |
| "learning_rate": 2.840909090909091e-07, | |
| "loss": 0.9168, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.019886363636363636, | |
| "grad_norm": 5.471309616694239, | |
| "learning_rate": 3.409090909090909e-07, | |
| "loss": 0.9448, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.022727272727272728, | |
| "grad_norm": 4.958288778905564, | |
| "learning_rate": 3.9772727272727276e-07, | |
| "loss": 0.8793, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.02556818181818182, | |
| "grad_norm": 4.933702183083632, | |
| "learning_rate": 4.5454545454545457e-07, | |
| "loss": 0.8895, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.028409090909090908, | |
| "grad_norm": 5.2766663694969616, | |
| "learning_rate": 5.113636363636364e-07, | |
| "loss": 0.9382, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03125, | |
| "grad_norm": 5.038528653630425, | |
| "learning_rate": 5.681818181818182e-07, | |
| "loss": 0.9115, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.03409090909090909, | |
| "grad_norm": 5.232611689675317, | |
| "learning_rate": 6.25e-07, | |
| "loss": 0.9292, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.036931818181818184, | |
| "grad_norm": 4.988139455338321, | |
| "learning_rate": 6.818181818181818e-07, | |
| "loss": 0.9014, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.03977272727272727, | |
| "grad_norm": 4.309132705290965, | |
| "learning_rate": 7.386363636363638e-07, | |
| "loss": 0.8873, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.04261363636363636, | |
| "grad_norm": 4.307584281122824, | |
| "learning_rate": 7.954545454545455e-07, | |
| "loss": 0.8333, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.045454545454545456, | |
| "grad_norm": 4.26307165025195, | |
| "learning_rate": 8.522727272727273e-07, | |
| "loss": 0.9033, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.048295454545454544, | |
| "grad_norm": 4.39034120717091, | |
| "learning_rate": 9.090909090909091e-07, | |
| "loss": 0.8635, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.05113636363636364, | |
| "grad_norm": 3.84235956788758, | |
| "learning_rate": 9.65909090909091e-07, | |
| "loss": 0.8446, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.05397727272727273, | |
| "grad_norm": 3.6266638353706115, | |
| "learning_rate": 1.0227272727272729e-06, | |
| "loss": 0.8415, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.056818181818181816, | |
| "grad_norm": 3.476359615290905, | |
| "learning_rate": 1.0795454545454546e-06, | |
| "loss": 0.8419, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.05965909090909091, | |
| "grad_norm": 3.156716131666368, | |
| "learning_rate": 1.1363636363636364e-06, | |
| "loss": 0.8403, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.0625, | |
| "grad_norm": 2.9151648586315138, | |
| "learning_rate": 1.1931818181818183e-06, | |
| "loss": 0.8146, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.06534090909090909, | |
| "grad_norm": 2.6691014608856354, | |
| "learning_rate": 1.25e-06, | |
| "loss": 0.7736, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.06818181818181818, | |
| "grad_norm": 2.573471865342396, | |
| "learning_rate": 1.3068181818181819e-06, | |
| "loss": 0.805, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.07102272727272728, | |
| "grad_norm": 2.478042571741361, | |
| "learning_rate": 1.3636363636363636e-06, | |
| "loss": 0.7884, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.07386363636363637, | |
| "grad_norm": 2.681597923455157, | |
| "learning_rate": 1.4204545454545458e-06, | |
| "loss": 0.7906, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.07670454545454546, | |
| "grad_norm": 3.5128502346347874, | |
| "learning_rate": 1.4772727272727275e-06, | |
| "loss": 0.8005, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.07954545454545454, | |
| "grad_norm": 3.0997314079873997, | |
| "learning_rate": 1.5340909090909093e-06, | |
| "loss": 0.7681, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.08238636363636363, | |
| "grad_norm": 3.301016199914052, | |
| "learning_rate": 1.590909090909091e-06, | |
| "loss": 0.7411, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.08522727272727272, | |
| "grad_norm": 3.0132955912093293, | |
| "learning_rate": 1.6477272727272728e-06, | |
| "loss": 0.7587, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.08806818181818182, | |
| "grad_norm": 2.389542355785432, | |
| "learning_rate": 1.7045454545454546e-06, | |
| "loss": 0.7115, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.09090909090909091, | |
| "grad_norm": 2.651871278037176, | |
| "learning_rate": 1.7613636363636365e-06, | |
| "loss": 0.7563, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.09375, | |
| "grad_norm": 2.164829082034884, | |
| "learning_rate": 1.8181818181818183e-06, | |
| "loss": 0.7599, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.09659090909090909, | |
| "grad_norm": 1.9014357867340361, | |
| "learning_rate": 1.8750000000000003e-06, | |
| "loss": 0.6873, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.09943181818181818, | |
| "grad_norm": 1.7872751567008667, | |
| "learning_rate": 1.931818181818182e-06, | |
| "loss": 0.7364, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.10227272727272728, | |
| "grad_norm": 2.0723837790108677, | |
| "learning_rate": 1.9886363636363638e-06, | |
| "loss": 0.6644, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.10511363636363637, | |
| "grad_norm": 2.1307866633916137, | |
| "learning_rate": 2.0454545454545457e-06, | |
| "loss": 0.6943, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.10795454545454546, | |
| "grad_norm": 2.2707754760571093, | |
| "learning_rate": 2.1022727272727277e-06, | |
| "loss": 0.6941, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.11079545454545454, | |
| "grad_norm": 2.458749073452043, | |
| "learning_rate": 2.1590909090909092e-06, | |
| "loss": 0.6788, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.11363636363636363, | |
| "grad_norm": 2.042545090231337, | |
| "learning_rate": 2.2159090909090912e-06, | |
| "loss": 0.6434, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.11647727272727272, | |
| "grad_norm": 1.8580077128269024, | |
| "learning_rate": 2.2727272727272728e-06, | |
| "loss": 0.6573, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.11931818181818182, | |
| "grad_norm": 1.8639374232639823, | |
| "learning_rate": 2.3295454545454547e-06, | |
| "loss": 0.6623, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.12215909090909091, | |
| "grad_norm": 1.7743666280182162, | |
| "learning_rate": 2.3863636363636367e-06, | |
| "loss": 0.6453, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.125, | |
| "grad_norm": 1.7689256446618578, | |
| "learning_rate": 2.4431818181818182e-06, | |
| "loss": 0.6379, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.1278409090909091, | |
| "grad_norm": 1.763969205340361, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.6751, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.13068181818181818, | |
| "grad_norm": 2.0371464953854543, | |
| "learning_rate": 2.556818181818182e-06, | |
| "loss": 0.6405, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.13352272727272727, | |
| "grad_norm": 1.7777840033046852, | |
| "learning_rate": 2.6136363636363637e-06, | |
| "loss": 0.6252, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.13636363636363635, | |
| "grad_norm": 1.7117349580502965, | |
| "learning_rate": 2.6704545454545457e-06, | |
| "loss": 0.6364, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.13920454545454544, | |
| "grad_norm": 1.6579980263734242, | |
| "learning_rate": 2.7272727272727272e-06, | |
| "loss": 0.6876, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.14204545454545456, | |
| "grad_norm": 1.645561441261347, | |
| "learning_rate": 2.784090909090909e-06, | |
| "loss": 0.6562, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.14488636363636365, | |
| "grad_norm": 1.6373402368183252, | |
| "learning_rate": 2.8409090909090916e-06, | |
| "loss": 0.5826, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.14772727272727273, | |
| "grad_norm": 1.6128855461127896, | |
| "learning_rate": 2.897727272727273e-06, | |
| "loss": 0.6826, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.15056818181818182, | |
| "grad_norm": 1.640865236871538, | |
| "learning_rate": 2.954545454545455e-06, | |
| "loss": 0.7155, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.1534090909090909, | |
| "grad_norm": 1.4720357466178808, | |
| "learning_rate": 3.0113636363636366e-06, | |
| "loss": 0.6204, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.15625, | |
| "grad_norm": 1.472120144942263, | |
| "learning_rate": 3.0681818181818186e-06, | |
| "loss": 0.6389, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.1590909090909091, | |
| "grad_norm": 1.449589289015899, | |
| "learning_rate": 3.125e-06, | |
| "loss": 0.5947, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.16193181818181818, | |
| "grad_norm": 1.5010999353034318, | |
| "learning_rate": 3.181818181818182e-06, | |
| "loss": 0.621, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.16477272727272727, | |
| "grad_norm": 1.8187567778405078, | |
| "learning_rate": 3.2386363636363637e-06, | |
| "loss": 0.6318, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.16761363636363635, | |
| "grad_norm": 1.6541144238811112, | |
| "learning_rate": 3.2954545454545456e-06, | |
| "loss": 0.5619, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.17045454545454544, | |
| "grad_norm": 1.6963775548673772, | |
| "learning_rate": 3.352272727272727e-06, | |
| "loss": 0.5823, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.17329545454545456, | |
| "grad_norm": 1.4964288240432566, | |
| "learning_rate": 3.409090909090909e-06, | |
| "loss": 0.5662, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.17613636363636365, | |
| "grad_norm": 1.6832346234487932, | |
| "learning_rate": 3.4659090909090915e-06, | |
| "loss": 0.6159, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.17897727272727273, | |
| "grad_norm": 1.5799422346591692, | |
| "learning_rate": 3.522727272727273e-06, | |
| "loss": 0.6365, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.18181818181818182, | |
| "grad_norm": 1.6582299717205322, | |
| "learning_rate": 3.579545454545455e-06, | |
| "loss": 0.647, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.1846590909090909, | |
| "grad_norm": 1.800280271805853, | |
| "learning_rate": 3.6363636363636366e-06, | |
| "loss": 0.6799, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.1875, | |
| "grad_norm": 1.7351635081451664, | |
| "learning_rate": 3.6931818181818186e-06, | |
| "loss": 0.6107, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.1903409090909091, | |
| "grad_norm": 1.5238122131962493, | |
| "learning_rate": 3.7500000000000005e-06, | |
| "loss": 0.6242, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.19318181818181818, | |
| "grad_norm": 1.6460555069181215, | |
| "learning_rate": 3.806818181818182e-06, | |
| "loss": 0.6193, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.19602272727272727, | |
| "grad_norm": 1.6136927208782876, | |
| "learning_rate": 3.863636363636364e-06, | |
| "loss": 0.6163, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.19886363636363635, | |
| "grad_norm": 1.6283580159831965, | |
| "learning_rate": 3.9204545454545456e-06, | |
| "loss": 0.5589, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.20170454545454544, | |
| "grad_norm": 1.6416288138229345, | |
| "learning_rate": 3.9772727272727275e-06, | |
| "loss": 0.5439, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.20454545454545456, | |
| "grad_norm": 1.7429357324390875, | |
| "learning_rate": 4.0340909090909095e-06, | |
| "loss": 0.647, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.20738636363636365, | |
| "grad_norm": 1.491759374118448, | |
| "learning_rate": 4.0909090909090915e-06, | |
| "loss": 0.5954, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.21022727272727273, | |
| "grad_norm": 1.5655918386388736, | |
| "learning_rate": 4.1477272727272734e-06, | |
| "loss": 0.5902, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.21306818181818182, | |
| "grad_norm": 1.6576527986240395, | |
| "learning_rate": 4.204545454545455e-06, | |
| "loss": 0.5856, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.2159090909090909, | |
| "grad_norm": 1.6628001520235092, | |
| "learning_rate": 4.2613636363636365e-06, | |
| "loss": 0.5808, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.21875, | |
| "grad_norm": 1.6620023765271195, | |
| "learning_rate": 4.3181818181818185e-06, | |
| "loss": 0.5992, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.2215909090909091, | |
| "grad_norm": 1.4886709042670327, | |
| "learning_rate": 4.3750000000000005e-06, | |
| "loss": 0.6091, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.22443181818181818, | |
| "grad_norm": 1.4251511789149547, | |
| "learning_rate": 4.4318181818181824e-06, | |
| "loss": 0.5867, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.22727272727272727, | |
| "grad_norm": 1.5633386386066466, | |
| "learning_rate": 4.4886363636363636e-06, | |
| "loss": 0.591, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.23011363636363635, | |
| "grad_norm": 1.5634040607472635, | |
| "learning_rate": 4.5454545454545455e-06, | |
| "loss": 0.5265, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.23295454545454544, | |
| "grad_norm": 1.6229601346591955, | |
| "learning_rate": 4.6022727272727275e-06, | |
| "loss": 0.5428, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.23579545454545456, | |
| "grad_norm": 1.5566532175674166, | |
| "learning_rate": 4.6590909090909095e-06, | |
| "loss": 0.5638, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.23863636363636365, | |
| "grad_norm": 1.593543335185521, | |
| "learning_rate": 4.715909090909091e-06, | |
| "loss": 0.5767, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.24147727272727273, | |
| "grad_norm": 1.449398786815831, | |
| "learning_rate": 4.772727272727273e-06, | |
| "loss": 0.5628, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.24431818181818182, | |
| "grad_norm": 1.5387043103946096, | |
| "learning_rate": 4.829545454545455e-06, | |
| "loss": 0.5735, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.2471590909090909, | |
| "grad_norm": 1.6763905573278226, | |
| "learning_rate": 4.8863636363636365e-06, | |
| "loss": 0.6032, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 1.5583409873897445, | |
| "learning_rate": 4.9431818181818184e-06, | |
| "loss": 0.623, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.2528409090909091, | |
| "grad_norm": 1.6187547313549033, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5804, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.2556818181818182, | |
| "grad_norm": 1.4963976477876901, | |
| "learning_rate": 5.056818181818182e-06, | |
| "loss": 0.5697, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2585227272727273, | |
| "grad_norm": 1.5131795004333568, | |
| "learning_rate": 5.113636363636364e-06, | |
| "loss": 0.5565, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.26136363636363635, | |
| "grad_norm": 1.5548395196775626, | |
| "learning_rate": 5.170454545454546e-06, | |
| "loss": 0.5461, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.26420454545454547, | |
| "grad_norm": 1.552959319249568, | |
| "learning_rate": 5.2272727272727274e-06, | |
| "loss": 0.5643, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.26704545454545453, | |
| "grad_norm": 1.6047709356284223, | |
| "learning_rate": 5.28409090909091e-06, | |
| "loss": 0.5798, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.26988636363636365, | |
| "grad_norm": 1.948509778915627, | |
| "learning_rate": 5.340909090909091e-06, | |
| "loss": 0.5327, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.2727272727272727, | |
| "grad_norm": 1.5318432235348463, | |
| "learning_rate": 5.397727272727273e-06, | |
| "loss": 0.5541, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.2755681818181818, | |
| "grad_norm": 1.5208133763810194, | |
| "learning_rate": 5.4545454545454545e-06, | |
| "loss": 0.529, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.2784090909090909, | |
| "grad_norm": 1.6124894417629865, | |
| "learning_rate": 5.511363636363637e-06, | |
| "loss": 0.5464, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.28125, | |
| "grad_norm": 1.7110965153428093, | |
| "learning_rate": 5.568181818181818e-06, | |
| "loss": 0.534, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.2840909090909091, | |
| "grad_norm": 1.658571452279957, | |
| "learning_rate": 5.625e-06, | |
| "loss": 0.6114, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2869318181818182, | |
| "grad_norm": 1.47658441902133, | |
| "learning_rate": 5.681818181818183e-06, | |
| "loss": 0.5547, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.2897727272727273, | |
| "grad_norm": 1.5722026271709708, | |
| "learning_rate": 5.738636363636364e-06, | |
| "loss": 0.5702, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.29261363636363635, | |
| "grad_norm": 1.6209365630457135, | |
| "learning_rate": 5.795454545454546e-06, | |
| "loss": 0.5728, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.29545454545454547, | |
| "grad_norm": 1.6904389396003674, | |
| "learning_rate": 5.852272727272727e-06, | |
| "loss": 0.5769, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.29829545454545453, | |
| "grad_norm": 1.5294499189980586, | |
| "learning_rate": 5.90909090909091e-06, | |
| "loss": 0.5656, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.30113636363636365, | |
| "grad_norm": 1.5916214451637154, | |
| "learning_rate": 5.965909090909091e-06, | |
| "loss": 0.5088, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.3039772727272727, | |
| "grad_norm": 1.6347827282817056, | |
| "learning_rate": 6.022727272727273e-06, | |
| "loss": 0.5828, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.3068181818181818, | |
| "grad_norm": 1.5618332724517157, | |
| "learning_rate": 6.079545454545454e-06, | |
| "loss": 0.5231, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.3096590909090909, | |
| "grad_norm": 1.6485262715983184, | |
| "learning_rate": 6.136363636363637e-06, | |
| "loss": 0.538, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.3125, | |
| "grad_norm": 1.7278179784286467, | |
| "learning_rate": 6.193181818181818e-06, | |
| "loss": 0.5431, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.3153409090909091, | |
| "grad_norm": 1.4942725054179251, | |
| "learning_rate": 6.25e-06, | |
| "loss": 0.5042, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.3181818181818182, | |
| "grad_norm": 1.6979365186765227, | |
| "learning_rate": 6.306818181818183e-06, | |
| "loss": 0.5677, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.3210227272727273, | |
| "grad_norm": 1.8418176894538365, | |
| "learning_rate": 6.363636363636364e-06, | |
| "loss": 0.588, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.32386363636363635, | |
| "grad_norm": 1.640630991484406, | |
| "learning_rate": 6.420454545454546e-06, | |
| "loss": 0.5447, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.32670454545454547, | |
| "grad_norm": 1.6066554835877627, | |
| "learning_rate": 6.477272727272727e-06, | |
| "loss": 0.581, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.32954545454545453, | |
| "grad_norm": 1.8057754828363723, | |
| "learning_rate": 6.53409090909091e-06, | |
| "loss": 0.5621, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.33238636363636365, | |
| "grad_norm": 1.711754276144589, | |
| "learning_rate": 6.590909090909091e-06, | |
| "loss": 0.5918, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.3352272727272727, | |
| "grad_norm": 1.517896860756366, | |
| "learning_rate": 6.647727272727273e-06, | |
| "loss": 0.5693, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.3380681818181818, | |
| "grad_norm": 1.5312310643773122, | |
| "learning_rate": 6.704545454545454e-06, | |
| "loss": 0.5473, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.3409090909090909, | |
| "grad_norm": 1.6182714991455485, | |
| "learning_rate": 6.761363636363637e-06, | |
| "loss": 0.5566, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.34375, | |
| "grad_norm": 1.6852771913947449, | |
| "learning_rate": 6.818181818181818e-06, | |
| "loss": 0.5547, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.3465909090909091, | |
| "grad_norm": 1.5897680070295204, | |
| "learning_rate": 6.875e-06, | |
| "loss": 0.5691, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.3494318181818182, | |
| "grad_norm": 1.7417266228377517, | |
| "learning_rate": 6.931818181818183e-06, | |
| "loss": 0.5362, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.3522727272727273, | |
| "grad_norm": 1.5989703470783707, | |
| "learning_rate": 6.988636363636364e-06, | |
| "loss": 0.5282, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.35511363636363635, | |
| "grad_norm": 1.6580595085217762, | |
| "learning_rate": 7.045454545454546e-06, | |
| "loss": 0.4953, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.35795454545454547, | |
| "grad_norm": 1.689844091653614, | |
| "learning_rate": 7.102272727272727e-06, | |
| "loss": 0.5467, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.36079545454545453, | |
| "grad_norm": 1.8347469318031917, | |
| "learning_rate": 7.15909090909091e-06, | |
| "loss": 0.536, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.36363636363636365, | |
| "grad_norm": 1.5529169597119448, | |
| "learning_rate": 7.215909090909091e-06, | |
| "loss": 0.524, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.3664772727272727, | |
| "grad_norm": 1.6645761334595488, | |
| "learning_rate": 7.272727272727273e-06, | |
| "loss": 0.5332, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.3693181818181818, | |
| "grad_norm": 1.7305662363325451, | |
| "learning_rate": 7.329545454545455e-06, | |
| "loss": 0.5219, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3721590909090909, | |
| "grad_norm": 1.5405083703593692, | |
| "learning_rate": 7.386363636363637e-06, | |
| "loss": 0.503, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.375, | |
| "grad_norm": 1.5865801365534873, | |
| "learning_rate": 7.443181818181818e-06, | |
| "loss": 0.5625, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.3778409090909091, | |
| "grad_norm": 1.782393408869917, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 0.5186, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.3806818181818182, | |
| "grad_norm": 1.8819124061481232, | |
| "learning_rate": 7.556818181818183e-06, | |
| "loss": 0.5932, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.3835227272727273, | |
| "grad_norm": 1.614587862077087, | |
| "learning_rate": 7.613636363636364e-06, | |
| "loss": 0.5156, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.38636363636363635, | |
| "grad_norm": 1.6896450400418643, | |
| "learning_rate": 7.670454545454547e-06, | |
| "loss": 0.5428, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.38920454545454547, | |
| "grad_norm": 1.648603590044125, | |
| "learning_rate": 7.727272727272727e-06, | |
| "loss": 0.5574, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.39204545454545453, | |
| "grad_norm": 1.686987993970066, | |
| "learning_rate": 7.784090909090911e-06, | |
| "loss": 0.4862, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.39488636363636365, | |
| "grad_norm": 1.7918189069952541, | |
| "learning_rate": 7.840909090909091e-06, | |
| "loss": 0.5527, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.3977272727272727, | |
| "grad_norm": 1.733609620505762, | |
| "learning_rate": 7.897727272727273e-06, | |
| "loss": 0.5896, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.4005681818181818, | |
| "grad_norm": 1.7767681019689756, | |
| "learning_rate": 7.954545454545455e-06, | |
| "loss": 0.5337, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.4034090909090909, | |
| "grad_norm": 1.62706994093177, | |
| "learning_rate": 8.011363636363637e-06, | |
| "loss": 0.529, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.40625, | |
| "grad_norm": 1.6190932829360436, | |
| "learning_rate": 8.068181818181819e-06, | |
| "loss": 0.5471, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.4090909090909091, | |
| "grad_norm": 1.4581793842750983, | |
| "learning_rate": 8.125000000000001e-06, | |
| "loss": 0.5353, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.4119318181818182, | |
| "grad_norm": 1.5106234263427771, | |
| "learning_rate": 8.181818181818183e-06, | |
| "loss": 0.5934, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.4147727272727273, | |
| "grad_norm": 1.5456174591730625, | |
| "learning_rate": 8.238636363636365e-06, | |
| "loss": 0.5055, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.41761363636363635, | |
| "grad_norm": 1.5169605544770564, | |
| "learning_rate": 8.295454545454547e-06, | |
| "loss": 0.5273, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.42045454545454547, | |
| "grad_norm": 1.6894139710313294, | |
| "learning_rate": 8.352272727272727e-06, | |
| "loss": 0.4984, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.42329545454545453, | |
| "grad_norm": 1.447004877728819, | |
| "learning_rate": 8.40909090909091e-06, | |
| "loss": 0.5616, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.42613636363636365, | |
| "grad_norm": 1.4673916601494148, | |
| "learning_rate": 8.465909090909091e-06, | |
| "loss": 0.5056, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.4289772727272727, | |
| "grad_norm": 1.524122665837613, | |
| "learning_rate": 8.522727272727273e-06, | |
| "loss": 0.5458, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.4318181818181818, | |
| "grad_norm": 1.6973227539753544, | |
| "learning_rate": 8.579545454545455e-06, | |
| "loss": 0.4954, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.4346590909090909, | |
| "grad_norm": 1.6229879362305883, | |
| "learning_rate": 8.636363636363637e-06, | |
| "loss": 0.531, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.4375, | |
| "grad_norm": 1.6650450560733285, | |
| "learning_rate": 8.693181818181819e-06, | |
| "loss": 0.4925, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.4403409090909091, | |
| "grad_norm": 1.52976091657608, | |
| "learning_rate": 8.750000000000001e-06, | |
| "loss": 0.4879, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.4431818181818182, | |
| "grad_norm": 1.5124703795369776, | |
| "learning_rate": 8.806818181818183e-06, | |
| "loss": 0.5203, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.4460227272727273, | |
| "grad_norm": 1.5985625351303485, | |
| "learning_rate": 8.863636363636365e-06, | |
| "loss": 0.6027, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.44886363636363635, | |
| "grad_norm": 1.588345112852916, | |
| "learning_rate": 8.920454545454547e-06, | |
| "loss": 0.5724, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.45170454545454547, | |
| "grad_norm": 1.7377485705572997, | |
| "learning_rate": 8.977272727272727e-06, | |
| "loss": 0.5222, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.45454545454545453, | |
| "grad_norm": 1.7413287692817025, | |
| "learning_rate": 9.03409090909091e-06, | |
| "loss": 0.5237, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.45738636363636365, | |
| "grad_norm": 1.7851847389488553, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 0.5663, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.4602272727272727, | |
| "grad_norm": 1.6153479555325219, | |
| "learning_rate": 9.147727272727273e-06, | |
| "loss": 0.5366, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.4630681818181818, | |
| "grad_norm": 1.637251170315082, | |
| "learning_rate": 9.204545454545455e-06, | |
| "loss": 0.5531, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.4659090909090909, | |
| "grad_norm": 1.6749216054319256, | |
| "learning_rate": 9.261363636363637e-06, | |
| "loss": 0.5513, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.46875, | |
| "grad_norm": 1.762177071370672, | |
| "learning_rate": 9.318181818181819e-06, | |
| "loss": 0.5164, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.4715909090909091, | |
| "grad_norm": 1.5634911246597942, | |
| "learning_rate": 9.375000000000001e-06, | |
| "loss": 0.4601, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.4744318181818182, | |
| "grad_norm": 1.5530286654821466, | |
| "learning_rate": 9.431818181818183e-06, | |
| "loss": 0.5434, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.4772727272727273, | |
| "grad_norm": 1.4890754991029818, | |
| "learning_rate": 9.488636363636365e-06, | |
| "loss": 0.5621, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.48011363636363635, | |
| "grad_norm": 1.545653087393646, | |
| "learning_rate": 9.545454545454547e-06, | |
| "loss": 0.573, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.48295454545454547, | |
| "grad_norm": 1.6628595064300078, | |
| "learning_rate": 9.602272727272727e-06, | |
| "loss": 0.5293, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.48579545454545453, | |
| "grad_norm": 1.6199990435876728, | |
| "learning_rate": 9.65909090909091e-06, | |
| "loss": 0.5347, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.48863636363636365, | |
| "grad_norm": 1.7032779971266183, | |
| "learning_rate": 9.715909090909091e-06, | |
| "loss": 0.5563, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.4914772727272727, | |
| "grad_norm": 1.5057928727458723, | |
| "learning_rate": 9.772727272727273e-06, | |
| "loss": 0.5441, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.4943181818181818, | |
| "grad_norm": 1.7930017997140106, | |
| "learning_rate": 9.829545454545455e-06, | |
| "loss": 0.5041, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.4971590909090909, | |
| "grad_norm": 1.395243731703088, | |
| "learning_rate": 9.886363636363637e-06, | |
| "loss": 0.5926, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.617882984644789, | |
| "learning_rate": 9.943181818181819e-06, | |
| "loss": 0.5188, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.5028409090909091, | |
| "grad_norm": 1.6266270332812511, | |
| "learning_rate": 1e-05, | |
| "loss": 0.5424, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.5056818181818182, | |
| "grad_norm": 1.7283548702418214, | |
| "learning_rate": 9.999990166021438e-06, | |
| "loss": 0.5037, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.5085227272727273, | |
| "grad_norm": 1.5500412944724196, | |
| "learning_rate": 9.999960664124435e-06, | |
| "loss": 0.5191, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.5113636363636364, | |
| "grad_norm": 1.650388352182727, | |
| "learning_rate": 9.999911494425041e-06, | |
| "loss": 0.5709, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.5142045454545454, | |
| "grad_norm": 1.7425734419236485, | |
| "learning_rate": 9.999842657116667e-06, | |
| "loss": 0.5537, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.5170454545454546, | |
| "grad_norm": 1.7441409412425497, | |
| "learning_rate": 9.99975415247009e-06, | |
| "loss": 0.5236, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.5198863636363636, | |
| "grad_norm": 1.7268326575282316, | |
| "learning_rate": 9.999645980833454e-06, | |
| "loss": 0.5273, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.5227272727272727, | |
| "grad_norm": 1.5872940055299185, | |
| "learning_rate": 9.999518142632263e-06, | |
| "loss": 0.5294, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.5255681818181818, | |
| "grad_norm": 1.5813034283378942, | |
| "learning_rate": 9.999370638369377e-06, | |
| "loss": 0.5202, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.5284090909090909, | |
| "grad_norm": 1.7024661321437708, | |
| "learning_rate": 9.999203468625017e-06, | |
| "loss": 0.514, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.53125, | |
| "grad_norm": 1.670438625550961, | |
| "learning_rate": 9.999016634056764e-06, | |
| "loss": 0.5139, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.5340909090909091, | |
| "grad_norm": 1.669791599161175, | |
| "learning_rate": 9.998810135399545e-06, | |
| "loss": 0.5218, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.5369318181818182, | |
| "grad_norm": 1.644186005465554, | |
| "learning_rate": 9.998583973465647e-06, | |
| "loss": 0.5487, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.5397727272727273, | |
| "grad_norm": 1.5862364549108896, | |
| "learning_rate": 9.998338149144693e-06, | |
| "loss": 0.531, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.5426136363636364, | |
| "grad_norm": 1.5651823720767404, | |
| "learning_rate": 9.998072663403657e-06, | |
| "loss": 0.5238, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.5454545454545454, | |
| "grad_norm": 1.449366653938261, | |
| "learning_rate": 9.997787517286852e-06, | |
| "loss": 0.5527, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.5482954545454546, | |
| "grad_norm": 1.552109203796676, | |
| "learning_rate": 9.997482711915926e-06, | |
| "loss": 0.5169, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.5511363636363636, | |
| "grad_norm": 1.5316780272692387, | |
| "learning_rate": 9.99715824848986e-06, | |
| "loss": 0.5419, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.5539772727272727, | |
| "grad_norm": 1.5561708694343155, | |
| "learning_rate": 9.99681412828496e-06, | |
| "loss": 0.5414, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.5568181818181818, | |
| "grad_norm": 1.6124118322811405, | |
| "learning_rate": 9.996450352654852e-06, | |
| "loss": 0.5193, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.5596590909090909, | |
| "grad_norm": 1.7208000153520733, | |
| "learning_rate": 9.996066923030484e-06, | |
| "loss": 0.5567, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.5625, | |
| "grad_norm": 1.5778279523696555, | |
| "learning_rate": 9.995663840920109e-06, | |
| "loss": 0.5053, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.5653409090909091, | |
| "grad_norm": 1.4855103243467769, | |
| "learning_rate": 9.99524110790929e-06, | |
| "loss": 0.5226, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.5681818181818182, | |
| "grad_norm": 1.667759826450969, | |
| "learning_rate": 9.994798725660884e-06, | |
| "loss": 0.4973, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5710227272727273, | |
| "grad_norm": 1.373799492473821, | |
| "learning_rate": 9.994336695915041e-06, | |
| "loss": 0.5007, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.5738636363636364, | |
| "grad_norm": 1.648039525222272, | |
| "learning_rate": 9.993855020489198e-06, | |
| "loss": 0.5144, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.5767045454545454, | |
| "grad_norm": 1.9647826400305184, | |
| "learning_rate": 9.993353701278072e-06, | |
| "loss": 0.543, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.5795454545454546, | |
| "grad_norm": 1.5693550891452794, | |
| "learning_rate": 9.992832740253646e-06, | |
| "loss": 0.5407, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.5823863636363636, | |
| "grad_norm": 1.6605237932745462, | |
| "learning_rate": 9.992292139465166e-06, | |
| "loss": 0.5447, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.5852272727272727, | |
| "grad_norm": 1.5688128583768695, | |
| "learning_rate": 9.991731901039137e-06, | |
| "loss": 0.5655, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.5880681818181818, | |
| "grad_norm": 1.6702105284432525, | |
| "learning_rate": 9.991152027179307e-06, | |
| "loss": 0.5074, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.5909090909090909, | |
| "grad_norm": 1.6991173010152403, | |
| "learning_rate": 9.990552520166664e-06, | |
| "loss": 0.54, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.59375, | |
| "grad_norm": 1.5436278737148996, | |
| "learning_rate": 9.989933382359423e-06, | |
| "loss": 0.5029, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.5965909090909091, | |
| "grad_norm": 1.6440577419199078, | |
| "learning_rate": 9.989294616193018e-06, | |
| "loss": 0.4773, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.5994318181818182, | |
| "grad_norm": 1.7071361354898558, | |
| "learning_rate": 9.988636224180097e-06, | |
| "loss": 0.5053, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.6022727272727273, | |
| "grad_norm": 1.7456629565862272, | |
| "learning_rate": 9.9879582089105e-06, | |
| "loss": 0.5004, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.6051136363636364, | |
| "grad_norm": 1.5909572264549439, | |
| "learning_rate": 9.987260573051268e-06, | |
| "loss": 0.5599, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.6079545454545454, | |
| "grad_norm": 1.7034637597618845, | |
| "learning_rate": 9.986543319346613e-06, | |
| "loss": 0.5436, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.6107954545454546, | |
| "grad_norm": 1.5069569616160918, | |
| "learning_rate": 9.985806450617916e-06, | |
| "loss": 0.5251, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.6136363636363636, | |
| "grad_norm": 1.5941088473148675, | |
| "learning_rate": 9.985049969763721e-06, | |
| "loss": 0.4738, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.6164772727272727, | |
| "grad_norm": 1.5111577587089249, | |
| "learning_rate": 9.984273879759713e-06, | |
| "loss": 0.5258, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.6193181818181818, | |
| "grad_norm": 1.4761221362965744, | |
| "learning_rate": 9.983478183658712e-06, | |
| "loss": 0.4785, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.6221590909090909, | |
| "grad_norm": 1.5040577229652037, | |
| "learning_rate": 9.982662884590662e-06, | |
| "loss": 0.4901, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 1.627974036727338, | |
| "learning_rate": 9.981827985762618e-06, | |
| "loss": 0.5002, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.6278409090909091, | |
| "grad_norm": 1.4751468229543794, | |
| "learning_rate": 9.980973490458728e-06, | |
| "loss": 0.5355, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.6306818181818182, | |
| "grad_norm": 1.5930987113010384, | |
| "learning_rate": 9.980099402040231e-06, | |
| "loss": 0.5668, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.6335227272727273, | |
| "grad_norm": 1.5813337316824179, | |
| "learning_rate": 9.97920572394543e-06, | |
| "loss": 0.4356, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.6363636363636364, | |
| "grad_norm": 1.774178789758969, | |
| "learning_rate": 9.978292459689692e-06, | |
| "loss": 0.5494, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.6392045454545454, | |
| "grad_norm": 1.5021051099138403, | |
| "learning_rate": 9.977359612865424e-06, | |
| "loss": 0.5346, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.6420454545454546, | |
| "grad_norm": 1.6931070839234672, | |
| "learning_rate": 9.976407187142066e-06, | |
| "loss": 0.5133, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.6448863636363636, | |
| "grad_norm": 1.725508341978023, | |
| "learning_rate": 9.975435186266069e-06, | |
| "loss": 0.5001, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.6477272727272727, | |
| "grad_norm": 1.4914218163125852, | |
| "learning_rate": 9.974443614060889e-06, | |
| "loss": 0.4941, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.6505681818181818, | |
| "grad_norm": 1.4782701475273787, | |
| "learning_rate": 9.973432474426968e-06, | |
| "loss": 0.5291, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.6534090909090909, | |
| "grad_norm": 1.5178330023195905, | |
| "learning_rate": 9.972401771341711e-06, | |
| "loss": 0.5118, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.65625, | |
| "grad_norm": 1.5738932861884023, | |
| "learning_rate": 9.971351508859488e-06, | |
| "loss": 0.5128, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.6590909090909091, | |
| "grad_norm": 1.5386697277681818, | |
| "learning_rate": 9.970281691111598e-06, | |
| "loss": 0.4767, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.6619318181818182, | |
| "grad_norm": 1.4630099429104129, | |
| "learning_rate": 9.969192322306271e-06, | |
| "loss": 0.4969, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.6647727272727273, | |
| "grad_norm": 1.3480903948948726, | |
| "learning_rate": 9.968083406728637e-06, | |
| "loss": 0.5732, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.6676136363636364, | |
| "grad_norm": 1.6632107802937945, | |
| "learning_rate": 9.966954948740717e-06, | |
| "loss": 0.4828, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.6704545454545454, | |
| "grad_norm": 1.479782256264717, | |
| "learning_rate": 9.965806952781402e-06, | |
| "loss": 0.5209, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.6732954545454546, | |
| "grad_norm": 1.6152559879763004, | |
| "learning_rate": 9.964639423366442e-06, | |
| "loss": 0.5247, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.6761363636363636, | |
| "grad_norm": 1.355213083704857, | |
| "learning_rate": 9.96345236508842e-06, | |
| "loss": 0.54, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.6789772727272727, | |
| "grad_norm": 1.4708340790563177, | |
| "learning_rate": 9.962245782616734e-06, | |
| "loss": 0.5178, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.6818181818181818, | |
| "grad_norm": 1.7002734228704037, | |
| "learning_rate": 9.961019680697593e-06, | |
| "loss": 0.4839, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.6846590909090909, | |
| "grad_norm": 1.5855366607994574, | |
| "learning_rate": 9.959774064153977e-06, | |
| "loss": 0.4664, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.6875, | |
| "grad_norm": 1.4710861883816675, | |
| "learning_rate": 9.958508937885633e-06, | |
| "loss": 0.5202, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.6903409090909091, | |
| "grad_norm": 1.5665588077627386, | |
| "learning_rate": 9.957224306869053e-06, | |
| "loss": 0.5082, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.6931818181818182, | |
| "grad_norm": 1.5137831929607448, | |
| "learning_rate": 9.955920176157448e-06, | |
| "loss": 0.4817, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.6960227272727273, | |
| "grad_norm": 1.5362365311838206, | |
| "learning_rate": 9.954596550880735e-06, | |
| "loss": 0.4949, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.6988636363636364, | |
| "grad_norm": 1.5434852584227288, | |
| "learning_rate": 9.953253436245519e-06, | |
| "loss": 0.5229, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.7017045454545454, | |
| "grad_norm": 1.4438612383868545, | |
| "learning_rate": 9.951890837535058e-06, | |
| "loss": 0.5273, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.7045454545454546, | |
| "grad_norm": 1.6229915207728822, | |
| "learning_rate": 9.950508760109265e-06, | |
| "loss": 0.4944, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.7073863636363636, | |
| "grad_norm": 1.3885376242229452, | |
| "learning_rate": 9.949107209404664e-06, | |
| "loss": 0.541, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.7102272727272727, | |
| "grad_norm": 1.4719200143551339, | |
| "learning_rate": 9.947686190934385e-06, | |
| "loss": 0.5062, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.7130681818181818, | |
| "grad_norm": 1.4430803001668344, | |
| "learning_rate": 9.946245710288132e-06, | |
| "loss": 0.5253, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.7159090909090909, | |
| "grad_norm": 1.4755305938799816, | |
| "learning_rate": 9.944785773132168e-06, | |
| "loss": 0.4852, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.71875, | |
| "grad_norm": 1.5136651249966744, | |
| "learning_rate": 9.94330638520929e-06, | |
| "loss": 0.5463, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.7215909090909091, | |
| "grad_norm": 1.5294777974746627, | |
| "learning_rate": 9.941807552338805e-06, | |
| "loss": 0.475, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.7244318181818182, | |
| "grad_norm": 1.605259878893428, | |
| "learning_rate": 9.940289280416509e-06, | |
| "loss": 0.4776, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.7272727272727273, | |
| "grad_norm": 1.315713602097482, | |
| "learning_rate": 9.938751575414663e-06, | |
| "loss": 0.5175, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.7301136363636364, | |
| "grad_norm": 1.3927615412027392, | |
| "learning_rate": 9.937194443381972e-06, | |
| "loss": 0.4765, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.7329545454545454, | |
| "grad_norm": 1.4339295292132817, | |
| "learning_rate": 9.935617890443557e-06, | |
| "loss": 0.4985, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.7357954545454546, | |
| "grad_norm": 1.3968961127996686, | |
| "learning_rate": 9.934021922800931e-06, | |
| "loss": 0.4704, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.7386363636363636, | |
| "grad_norm": 1.44031807612239, | |
| "learning_rate": 9.932406546731981e-06, | |
| "loss": 0.5186, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.7414772727272727, | |
| "grad_norm": 1.5831116349041292, | |
| "learning_rate": 9.930771768590934e-06, | |
| "loss": 0.4988, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.7443181818181818, | |
| "grad_norm": 1.371989349404777, | |
| "learning_rate": 9.929117594808341e-06, | |
| "loss": 0.5018, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.7471590909090909, | |
| "grad_norm": 1.5053430727565054, | |
| "learning_rate": 9.927444031891044e-06, | |
| "loss": 0.5703, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 2.0336405552487338, | |
| "learning_rate": 9.92575108642216e-06, | |
| "loss": 0.4771, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.7528409090909091, | |
| "grad_norm": 1.3896429508762578, | |
| "learning_rate": 9.924038765061042e-06, | |
| "loss": 0.4709, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.7556818181818182, | |
| "grad_norm": 1.5824500328291682, | |
| "learning_rate": 9.92230707454326e-06, | |
| "loss": 0.4818, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.7585227272727273, | |
| "grad_norm": 1.537798794932119, | |
| "learning_rate": 9.92055602168058e-06, | |
| "loss": 0.4845, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.7613636363636364, | |
| "grad_norm": 1.4323222954770605, | |
| "learning_rate": 9.918785613360931e-06, | |
| "loss": 0.5638, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.7642045454545454, | |
| "grad_norm": 1.5617485354631635, | |
| "learning_rate": 9.916995856548371e-06, | |
| "loss": 0.4595, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.7670454545454546, | |
| "grad_norm": 1.6295758621364878, | |
| "learning_rate": 9.915186758283072e-06, | |
| "loss": 0.5003, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.7698863636363636, | |
| "grad_norm": 1.7193205168854822, | |
| "learning_rate": 9.913358325681292e-06, | |
| "loss": 0.4921, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.7727272727272727, | |
| "grad_norm": 1.4343207005446186, | |
| "learning_rate": 9.911510565935335e-06, | |
| "loss": 0.5572, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.7755681818181818, | |
| "grad_norm": 1.5153801998849168, | |
| "learning_rate": 9.909643486313533e-06, | |
| "loss": 0.4599, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.7784090909090909, | |
| "grad_norm": 1.5255133825708138, | |
| "learning_rate": 9.907757094160217e-06, | |
| "loss": 0.4847, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.78125, | |
| "grad_norm": 1.5245888606968223, | |
| "learning_rate": 9.905851396895679e-06, | |
| "loss": 0.5101, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.7840909090909091, | |
| "grad_norm": 1.54702192325514, | |
| "learning_rate": 9.903926402016153e-06, | |
| "loss": 0.5076, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.7869318181818182, | |
| "grad_norm": 1.6313992214857218, | |
| "learning_rate": 9.901982117093786e-06, | |
| "loss": 0.4601, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.7897727272727273, | |
| "grad_norm": 1.5915121599329511, | |
| "learning_rate": 9.900018549776598e-06, | |
| "loss": 0.4865, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.7926136363636364, | |
| "grad_norm": 1.5326657935013723, | |
| "learning_rate": 9.898035707788462e-06, | |
| "loss": 0.5098, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.7954545454545454, | |
| "grad_norm": 1.701813376109124, | |
| "learning_rate": 9.896033598929069e-06, | |
| "loss": 0.5362, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.7982954545454546, | |
| "grad_norm": 1.4800001962097533, | |
| "learning_rate": 9.894012231073895e-06, | |
| "loss": 0.5324, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.8011363636363636, | |
| "grad_norm": 1.4483575554579828, | |
| "learning_rate": 9.891971612174176e-06, | |
| "loss": 0.4944, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.8039772727272727, | |
| "grad_norm": 1.480969574220171, | |
| "learning_rate": 9.889911750256873e-06, | |
| "loss": 0.514, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.8068181818181818, | |
| "grad_norm": 1.6355803470028214, | |
| "learning_rate": 9.88783265342464e-06, | |
| "loss": 0.5011, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.8096590909090909, | |
| "grad_norm": 1.4490799988047873, | |
| "learning_rate": 9.885734329855798e-06, | |
| "loss": 0.5329, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.8125, | |
| "grad_norm": 1.5125220233355698, | |
| "learning_rate": 9.883616787804292e-06, | |
| "loss": 0.5061, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.8153409090909091, | |
| "grad_norm": 1.5492210249492218, | |
| "learning_rate": 9.881480035599667e-06, | |
| "loss": 0.5153, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.8181818181818182, | |
| "grad_norm": 1.400547790361256, | |
| "learning_rate": 9.879324081647035e-06, | |
| "loss": 0.5208, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.8210227272727273, | |
| "grad_norm": 1.5862814840089992, | |
| "learning_rate": 9.877148934427037e-06, | |
| "loss": 0.5043, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.8238636363636364, | |
| "grad_norm": 1.5846145487408034, | |
| "learning_rate": 9.874954602495811e-06, | |
| "loss": 0.4685, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.8267045454545454, | |
| "grad_norm": 1.5029187410604994, | |
| "learning_rate": 9.872741094484965e-06, | |
| "loss": 0.469, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.8295454545454546, | |
| "grad_norm": 1.4366394515179206, | |
| "learning_rate": 9.870508419101536e-06, | |
| "loss": 0.5196, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.8323863636363636, | |
| "grad_norm": 1.6410923835471025, | |
| "learning_rate": 9.868256585127956e-06, | |
| "loss": 0.5072, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.8352272727272727, | |
| "grad_norm": 1.6058263511201358, | |
| "learning_rate": 9.865985601422018e-06, | |
| "loss": 0.5226, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.8380681818181818, | |
| "grad_norm": 1.5406427651119536, | |
| "learning_rate": 9.863695476916846e-06, | |
| "loss": 0.5513, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.8409090909090909, | |
| "grad_norm": 1.5483143683420586, | |
| "learning_rate": 9.861386220620853e-06, | |
| "loss": 0.5272, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.84375, | |
| "grad_norm": 1.5737342478379488, | |
| "learning_rate": 9.859057841617709e-06, | |
| "loss": 0.4981, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.8465909090909091, | |
| "grad_norm": 1.5552099127946464, | |
| "learning_rate": 9.856710349066307e-06, | |
| "loss": 0.5316, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.8494318181818182, | |
| "grad_norm": 1.3775976434027732, | |
| "learning_rate": 9.854343752200725e-06, | |
| "loss": 0.5028, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.8522727272727273, | |
| "grad_norm": 1.534882345079291, | |
| "learning_rate": 9.851958060330186e-06, | |
| "loss": 0.5037, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.8551136363636364, | |
| "grad_norm": 1.3965739101658592, | |
| "learning_rate": 9.849553282839025e-06, | |
| "loss": 0.4902, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.8579545454545454, | |
| "grad_norm": 1.5164723826063233, | |
| "learning_rate": 9.847129429186662e-06, | |
| "loss": 0.453, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.8607954545454546, | |
| "grad_norm": 1.4716195200624351, | |
| "learning_rate": 9.844686508907538e-06, | |
| "loss": 0.5214, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.8636363636363636, | |
| "grad_norm": 1.496443165367598, | |
| "learning_rate": 9.842224531611106e-06, | |
| "loss": 0.4291, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.8664772727272727, | |
| "grad_norm": 1.4268349606630073, | |
| "learning_rate": 9.839743506981783e-06, | |
| "loss": 0.5016, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.8693181818181818, | |
| "grad_norm": 1.552129131024092, | |
| "learning_rate": 9.8372434447789e-06, | |
| "loss": 0.499, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.8721590909090909, | |
| "grad_norm": 1.3993525226391348, | |
| "learning_rate": 9.834724354836684e-06, | |
| "loss": 0.5012, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.875, | |
| "grad_norm": 1.3911264167867943, | |
| "learning_rate": 9.832186247064205e-06, | |
| "loss": 0.5079, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.8778409090909091, | |
| "grad_norm": 1.4874539088146177, | |
| "learning_rate": 9.829629131445342e-06, | |
| "loss": 0.4655, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.8806818181818182, | |
| "grad_norm": 1.4484874853487535, | |
| "learning_rate": 9.827053018038743e-06, | |
| "loss": 0.447, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.8835227272727273, | |
| "grad_norm": 1.4169976323362639, | |
| "learning_rate": 9.824457916977785e-06, | |
| "loss": 0.5169, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.8863636363636364, | |
| "grad_norm": 1.5792992862826667, | |
| "learning_rate": 9.821843838470536e-06, | |
| "loss": 0.4499, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.8892045454545454, | |
| "grad_norm": 1.3935050163686387, | |
| "learning_rate": 9.819210792799711e-06, | |
| "loss": 0.4836, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.8920454545454546, | |
| "grad_norm": 1.560801731687215, | |
| "learning_rate": 9.816558790322638e-06, | |
| "loss": 0.4627, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.8948863636363636, | |
| "grad_norm": 1.8175856225601175, | |
| "learning_rate": 9.81388784147121e-06, | |
| "loss": 0.4842, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.8977272727272727, | |
| "grad_norm": 1.59068026179008, | |
| "learning_rate": 9.811197956751851e-06, | |
| "loss": 0.5668, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.9005681818181818, | |
| "grad_norm": 1.4641719221220586, | |
| "learning_rate": 9.808489146745466e-06, | |
| "loss": 0.4721, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.9034090909090909, | |
| "grad_norm": 1.433777355819498, | |
| "learning_rate": 9.805761422107407e-06, | |
| "loss": 0.4943, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.90625, | |
| "grad_norm": 1.5468925623745085, | |
| "learning_rate": 9.803014793567429e-06, | |
| "loss": 0.4907, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.9090909090909091, | |
| "grad_norm": 1.581396450996295, | |
| "learning_rate": 9.800249271929645e-06, | |
| "loss": 0.4494, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.9119318181818182, | |
| "grad_norm": 1.4624512226735435, | |
| "learning_rate": 9.797464868072489e-06, | |
| "loss": 0.4803, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.9147727272727273, | |
| "grad_norm": 1.431526222855722, | |
| "learning_rate": 9.794661592948665e-06, | |
| "loss": 0.4698, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.9176136363636364, | |
| "grad_norm": 1.561310299668803, | |
| "learning_rate": 9.791839457585118e-06, | |
| "loss": 0.5117, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.9204545454545454, | |
| "grad_norm": 1.625460220250549, | |
| "learning_rate": 9.788998473082968e-06, | |
| "loss": 0.4971, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.9232954545454546, | |
| "grad_norm": 1.6121422574836697, | |
| "learning_rate": 9.786138650617494e-06, | |
| "loss": 0.5365, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.9261363636363636, | |
| "grad_norm": 1.502479264528449, | |
| "learning_rate": 9.783260001438066e-06, | |
| "loss": 0.4695, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.9289772727272727, | |
| "grad_norm": 1.505956180933774, | |
| "learning_rate": 9.780362536868113e-06, | |
| "loss": 0.4844, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.9318181818181818, | |
| "grad_norm": 1.6301168834074384, | |
| "learning_rate": 9.777446268305079e-06, | |
| "loss": 0.4709, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.9346590909090909, | |
| "grad_norm": 1.5755579003886004, | |
| "learning_rate": 9.774511207220369e-06, | |
| "loss": 0.5236, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.9375, | |
| "grad_norm": 1.6499540937400865, | |
| "learning_rate": 9.77155736515932e-06, | |
| "loss": 0.5175, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.9403409090909091, | |
| "grad_norm": 1.5960557618257305, | |
| "learning_rate": 9.768584753741134e-06, | |
| "loss": 0.498, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.9431818181818182, | |
| "grad_norm": 1.5941569372324007, | |
| "learning_rate": 9.765593384658855e-06, | |
| "loss": 0.4732, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.9460227272727273, | |
| "grad_norm": 1.5228325929762483, | |
| "learning_rate": 9.762583269679304e-06, | |
| "loss": 0.4431, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.9488636363636364, | |
| "grad_norm": 1.490741833018044, | |
| "learning_rate": 9.759554420643043e-06, | |
| "loss": 0.5188, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.9517045454545454, | |
| "grad_norm": 1.6002668286658763, | |
| "learning_rate": 9.756506849464327e-06, | |
| "loss": 0.4839, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.9545454545454546, | |
| "grad_norm": 1.4598244710308395, | |
| "learning_rate": 9.753440568131056e-06, | |
| "loss": 0.4926, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.9573863636363636, | |
| "grad_norm": 1.4532786929990509, | |
| "learning_rate": 9.750355588704728e-06, | |
| "loss": 0.509, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.9602272727272727, | |
| "grad_norm": 1.6104333498186194, | |
| "learning_rate": 9.74725192332039e-06, | |
| "loss": 0.4843, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.9630681818181818, | |
| "grad_norm": 1.5091072452463585, | |
| "learning_rate": 9.744129584186599e-06, | |
| "loss": 0.5373, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.9659090909090909, | |
| "grad_norm": 1.3854787912992945, | |
| "learning_rate": 9.740988583585356e-06, | |
| "loss": 0.4462, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.96875, | |
| "grad_norm": 1.3149781820395967, | |
| "learning_rate": 9.737828933872076e-06, | |
| "loss": 0.5111, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.9715909090909091, | |
| "grad_norm": 1.4869043041061063, | |
| "learning_rate": 9.73465064747553e-06, | |
| "loss": 0.4843, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.9744318181818182, | |
| "grad_norm": 1.4416568690760034, | |
| "learning_rate": 9.731453736897796e-06, | |
| "loss": 0.4658, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.9772727272727273, | |
| "grad_norm": 1.5490848888116837, | |
| "learning_rate": 9.72823821471422e-06, | |
| "loss": 0.4676, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.9801136363636364, | |
| "grad_norm": 1.3810890147044554, | |
| "learning_rate": 9.725004093573343e-06, | |
| "loss": 0.495, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.9829545454545454, | |
| "grad_norm": 1.5581162695244781, | |
| "learning_rate": 9.721751386196885e-06, | |
| "loss": 0.5413, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.9857954545454546, | |
| "grad_norm": 1.5331178001377477, | |
| "learning_rate": 9.718480105379663e-06, | |
| "loss": 0.4913, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.9886363636363636, | |
| "grad_norm": 1.3200748312929138, | |
| "learning_rate": 9.715190263989562e-06, | |
| "loss": 0.471, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.9914772727272727, | |
| "grad_norm": 1.3932172431217436, | |
| "learning_rate": 9.711881874967471e-06, | |
| "loss": 0.4746, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.9943181818181818, | |
| "grad_norm": 1.596928748921533, | |
| "learning_rate": 9.708554951327243e-06, | |
| "loss": 0.496, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.9971590909090909, | |
| "grad_norm": 1.5042003988892965, | |
| "learning_rate": 9.705209506155635e-06, | |
| "loss": 0.5038, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 1.34623021375936, | |
| "learning_rate": 9.701845552612261e-06, | |
| "loss": 0.4584, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 1.0028409090909092, | |
| "grad_norm": 1.3935683216098373, | |
| "learning_rate": 9.698463103929542e-06, | |
| "loss": 0.416, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 1.0056818181818181, | |
| "grad_norm": 1.2572279971528715, | |
| "learning_rate": 9.695062173412648e-06, | |
| "loss": 0.3841, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 1.0085227272727273, | |
| "grad_norm": 1.331784432855466, | |
| "learning_rate": 9.69164277443945e-06, | |
| "loss": 0.411, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.0113636363636365, | |
| "grad_norm": 1.3861844220787067, | |
| "learning_rate": 9.688204920460467e-06, | |
| "loss": 0.4348, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 1.0142045454545454, | |
| "grad_norm": 1.3709086922612779, | |
| "learning_rate": 9.68474862499881e-06, | |
| "loss": 0.4243, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 1.0170454545454546, | |
| "grad_norm": 1.4954332771145582, | |
| "learning_rate": 9.681273901650134e-06, | |
| "loss": 0.3301, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 1.0198863636363635, | |
| "grad_norm": 1.5628318749217331, | |
| "learning_rate": 9.677780764082583e-06, | |
| "loss": 0.3853, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 1.0227272727272727, | |
| "grad_norm": 1.4017875812891134, | |
| "learning_rate": 9.67426922603673e-06, | |
| "loss": 0.4367, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.0255681818181819, | |
| "grad_norm": 1.5499407509600363, | |
| "learning_rate": 9.670739301325534e-06, | |
| "loss": 0.4372, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 1.0284090909090908, | |
| "grad_norm": 1.3221335169217352, | |
| "learning_rate": 9.667191003834276e-06, | |
| "loss": 0.3944, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 1.03125, | |
| "grad_norm": 1.2928319666303545, | |
| "learning_rate": 9.663624347520506e-06, | |
| "loss": 0.3913, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 1.0340909090909092, | |
| "grad_norm": 1.3237996792052562, | |
| "learning_rate": 9.660039346413994e-06, | |
| "loss": 0.3677, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.0369318181818181, | |
| "grad_norm": 1.356408795095918, | |
| "learning_rate": 9.65643601461667e-06, | |
| "loss": 0.3683, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.0397727272727273, | |
| "grad_norm": 1.3648543824708266, | |
| "learning_rate": 9.65281436630257e-06, | |
| "loss": 0.4299, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 1.0426136363636365, | |
| "grad_norm": 1.306204197525364, | |
| "learning_rate": 9.649174415717776e-06, | |
| "loss": 0.3814, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 1.0454545454545454, | |
| "grad_norm": 1.378237936287675, | |
| "learning_rate": 9.64551617718037e-06, | |
| "loss": 0.3856, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 1.0482954545454546, | |
| "grad_norm": 1.347725144403338, | |
| "learning_rate": 9.641839665080363e-06, | |
| "loss": 0.3868, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 1.0511363636363635, | |
| "grad_norm": 1.2789287048030384, | |
| "learning_rate": 9.638144893879657e-06, | |
| "loss": 0.4427, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.0539772727272727, | |
| "grad_norm": 1.4122383248910728, | |
| "learning_rate": 9.634431878111969e-06, | |
| "loss": 0.372, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 1.0568181818181819, | |
| "grad_norm": 1.382980088005404, | |
| "learning_rate": 9.630700632382787e-06, | |
| "loss": 0.398, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 1.0596590909090908, | |
| "grad_norm": 1.4228786699415132, | |
| "learning_rate": 9.626951171369306e-06, | |
| "loss": 0.4086, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 1.0625, | |
| "grad_norm": 1.2951379080398953, | |
| "learning_rate": 9.623183509820376e-06, | |
| "loss": 0.4102, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 1.0653409090909092, | |
| "grad_norm": 1.3372702622351302, | |
| "learning_rate": 9.619397662556434e-06, | |
| "loss": 0.387, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.0681818181818181, | |
| "grad_norm": 1.4961377119399946, | |
| "learning_rate": 9.61559364446946e-06, | |
| "loss": 0.3735, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 1.0710227272727273, | |
| "grad_norm": 1.4498025423293337, | |
| "learning_rate": 9.611771470522908e-06, | |
| "loss": 0.4034, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 1.0738636363636365, | |
| "grad_norm": 1.3626788600371367, | |
| "learning_rate": 9.607931155751646e-06, | |
| "loss": 0.321, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 1.0767045454545454, | |
| "grad_norm": 1.4483349711836415, | |
| "learning_rate": 9.604072715261902e-06, | |
| "loss": 0.3996, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 1.0795454545454546, | |
| "grad_norm": 1.4328248222967257, | |
| "learning_rate": 9.600196164231209e-06, | |
| "loss": 0.4367, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.0823863636363635, | |
| "grad_norm": 1.5892165025198344, | |
| "learning_rate": 9.596301517908329e-06, | |
| "loss": 0.3816, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 1.0852272727272727, | |
| "grad_norm": 1.4702034999304534, | |
| "learning_rate": 9.592388791613215e-06, | |
| "loss": 0.3685, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 1.0880681818181819, | |
| "grad_norm": 1.4201126245689717, | |
| "learning_rate": 9.588458000736929e-06, | |
| "loss": 0.4035, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 1.0909090909090908, | |
| "grad_norm": 1.4451056473536599, | |
| "learning_rate": 9.584509160741599e-06, | |
| "loss": 0.3979, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 1.09375, | |
| "grad_norm": 1.4239268094337205, | |
| "learning_rate": 9.580542287160348e-06, | |
| "loss": 0.4152, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.0965909090909092, | |
| "grad_norm": 1.3727180391928673, | |
| "learning_rate": 9.576557395597237e-06, | |
| "loss": 0.351, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 1.0994318181818181, | |
| "grad_norm": 1.3764065056655754, | |
| "learning_rate": 9.572554501727198e-06, | |
| "loss": 0.3993, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 1.1022727272727273, | |
| "grad_norm": 1.350149088431701, | |
| "learning_rate": 9.568533621295982e-06, | |
| "loss": 0.3802, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.1051136363636365, | |
| "grad_norm": 1.3494913492931555, | |
| "learning_rate": 9.564494770120089e-06, | |
| "loss": 0.3976, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 1.1079545454545454, | |
| "grad_norm": 1.2895978086734443, | |
| "learning_rate": 9.560437964086713e-06, | |
| "loss": 0.3737, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.1107954545454546, | |
| "grad_norm": 1.3580957096749768, | |
| "learning_rate": 9.556363219153664e-06, | |
| "loss": 0.4237, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 1.1136363636363635, | |
| "grad_norm": 1.448759353025191, | |
| "learning_rate": 9.55227055134933e-06, | |
| "loss": 0.365, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 1.1164772727272727, | |
| "grad_norm": 1.4090875794073592, | |
| "learning_rate": 9.548159976772593e-06, | |
| "loss": 0.3352, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 1.1193181818181819, | |
| "grad_norm": 1.318489892775961, | |
| "learning_rate": 9.544031511592772e-06, | |
| "loss": 0.3801, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 1.1221590909090908, | |
| "grad_norm": 1.482014367782798, | |
| "learning_rate": 9.539885172049563e-06, | |
| "loss": 0.4543, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.125, | |
| "grad_norm": 1.5905204206786512, | |
| "learning_rate": 9.535720974452973e-06, | |
| "loss": 0.3749, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 1.1278409090909092, | |
| "grad_norm": 1.4661858452157561, | |
| "learning_rate": 9.531538935183252e-06, | |
| "loss": 0.3671, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 1.1306818181818181, | |
| "grad_norm": 1.4153932607502078, | |
| "learning_rate": 9.527339070690833e-06, | |
| "loss": 0.4444, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 1.1335227272727273, | |
| "grad_norm": 1.2966945431847525, | |
| "learning_rate": 9.52312139749627e-06, | |
| "loss": 0.4391, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 1.1363636363636362, | |
| "grad_norm": 1.288707482023972, | |
| "learning_rate": 9.518885932190166e-06, | |
| "loss": 0.3696, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.1392045454545454, | |
| "grad_norm": 1.3246404279367519, | |
| "learning_rate": 9.514632691433108e-06, | |
| "loss": 0.3339, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 1.1420454545454546, | |
| "grad_norm": 1.274941281826731, | |
| "learning_rate": 9.510361691955607e-06, | |
| "loss": 0.3902, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 1.1448863636363638, | |
| "grad_norm": 1.2920978892355353, | |
| "learning_rate": 9.506072950558036e-06, | |
| "loss": 0.4159, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 1.1477272727272727, | |
| "grad_norm": 1.44258338460387, | |
| "learning_rate": 9.501766484110546e-06, | |
| "loss": 0.4162, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 1.1505681818181819, | |
| "grad_norm": 1.3556175583934928, | |
| "learning_rate": 9.497442309553017e-06, | |
| "loss": 0.4682, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.1534090909090908, | |
| "grad_norm": 1.4681216384949332, | |
| "learning_rate": 9.493100443894986e-06, | |
| "loss": 0.3963, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 1.15625, | |
| "grad_norm": 1.5449635342485868, | |
| "learning_rate": 9.488740904215578e-06, | |
| "loss": 0.3845, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 1.1590909090909092, | |
| "grad_norm": 1.4235183492577208, | |
| "learning_rate": 9.484363707663443e-06, | |
| "loss": 0.3596, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 1.1619318181818181, | |
| "grad_norm": 1.3972675999686617, | |
| "learning_rate": 9.47996887145668e-06, | |
| "loss": 0.3598, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 1.1647727272727273, | |
| "grad_norm": 1.3050109631890814, | |
| "learning_rate": 9.475556412882782e-06, | |
| "loss": 0.3975, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.1676136363636362, | |
| "grad_norm": 1.3019015711552302, | |
| "learning_rate": 9.471126349298557e-06, | |
| "loss": 0.3778, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 1.1704545454545454, | |
| "grad_norm": 1.3752118885869313, | |
| "learning_rate": 9.466678698130064e-06, | |
| "loss": 0.3948, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 1.1732954545454546, | |
| "grad_norm": 1.4120612953299163, | |
| "learning_rate": 9.46221347687255e-06, | |
| "loss": 0.4006, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 1.1761363636363638, | |
| "grad_norm": 1.3644262325212158, | |
| "learning_rate": 9.457730703090367e-06, | |
| "loss": 0.45, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 1.1789772727272727, | |
| "grad_norm": 1.4024791684682305, | |
| "learning_rate": 9.453230394416914e-06, | |
| "loss": 0.4321, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.1818181818181819, | |
| "grad_norm": 1.2518931861431282, | |
| "learning_rate": 9.448712568554572e-06, | |
| "loss": 0.4311, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 1.1846590909090908, | |
| "grad_norm": 1.3151977021290406, | |
| "learning_rate": 9.444177243274619e-06, | |
| "loss": 0.4405, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 1.1875, | |
| "grad_norm": 1.2870366213234796, | |
| "learning_rate": 9.43962443641717e-06, | |
| "loss": 0.3946, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 1.1903409090909092, | |
| "grad_norm": 1.2555358317281697, | |
| "learning_rate": 9.43505416589111e-06, | |
| "loss": 0.3904, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 1.1931818181818181, | |
| "grad_norm": 1.4290322380534968, | |
| "learning_rate": 9.430466449674014e-06, | |
| "loss": 0.4199, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.1960227272727273, | |
| "grad_norm": 1.3340923689952158, | |
| "learning_rate": 9.425861305812083e-06, | |
| "loss": 0.4016, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 1.1988636363636362, | |
| "grad_norm": 1.3506050543273165, | |
| "learning_rate": 9.421238752420075e-06, | |
| "loss": 0.3591, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 1.2017045454545454, | |
| "grad_norm": 1.3408737454429664, | |
| "learning_rate": 9.416598807681221e-06, | |
| "loss": 0.3958, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 1.2045454545454546, | |
| "grad_norm": 1.3066485596939463, | |
| "learning_rate": 9.411941489847172e-06, | |
| "loss": 0.3712, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 1.2073863636363638, | |
| "grad_norm": 1.386076628663598, | |
| "learning_rate": 9.40726681723791e-06, | |
| "loss": 0.363, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.2102272727272727, | |
| "grad_norm": 1.4001603892788743, | |
| "learning_rate": 9.402574808241692e-06, | |
| "loss": 0.3958, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 1.2130681818181819, | |
| "grad_norm": 1.5442221898856994, | |
| "learning_rate": 9.397865481314959e-06, | |
| "loss": 0.3936, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 1.2159090909090908, | |
| "grad_norm": 1.5030113636563918, | |
| "learning_rate": 9.393138854982283e-06, | |
| "loss": 0.4011, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 1.21875, | |
| "grad_norm": 1.437402205187118, | |
| "learning_rate": 9.388394947836278e-06, | |
| "loss": 0.4231, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 1.2215909090909092, | |
| "grad_norm": 1.395930365479985, | |
| "learning_rate": 9.38363377853754e-06, | |
| "loss": 0.4078, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.2244318181818181, | |
| "grad_norm": 1.4691665996144123, | |
| "learning_rate": 9.37885536581456e-06, | |
| "loss": 0.4, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 1.2272727272727273, | |
| "grad_norm": 1.3654497134259025, | |
| "learning_rate": 9.374059728463663e-06, | |
| "loss": 0.4093, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 1.2301136363636362, | |
| "grad_norm": 1.4295062613690053, | |
| "learning_rate": 9.369246885348926e-06, | |
| "loss": 0.384, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.2329545454545454, | |
| "grad_norm": 1.2387775411886048, | |
| "learning_rate": 9.364416855402107e-06, | |
| "loss": 0.3577, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 1.2357954545454546, | |
| "grad_norm": 1.3595503473660588, | |
| "learning_rate": 9.359569657622573e-06, | |
| "loss": 0.3957, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.2386363636363638, | |
| "grad_norm": 1.367378341936646, | |
| "learning_rate": 9.35470531107722e-06, | |
| "loss": 0.4374, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 1.2414772727272727, | |
| "grad_norm": 1.3007672389688, | |
| "learning_rate": 9.349823834900396e-06, | |
| "loss": 0.3906, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 1.2443181818181819, | |
| "grad_norm": 1.4358136508434372, | |
| "learning_rate": 9.344925248293837e-06, | |
| "loss": 0.3587, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.2471590909090908, | |
| "grad_norm": 1.3487005026003476, | |
| "learning_rate": 9.340009570526578e-06, | |
| "loss": 0.3957, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 1.3187450315078482, | |
| "learning_rate": 9.335076820934889e-06, | |
| "loss": 0.3913, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.2528409090909092, | |
| "grad_norm": 1.4046469997918478, | |
| "learning_rate": 9.330127018922195e-06, | |
| "loss": 0.3846, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 1.2556818181818181, | |
| "grad_norm": 1.3725821672416023, | |
| "learning_rate": 9.325160183958987e-06, | |
| "loss": 0.4077, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 1.2585227272727273, | |
| "grad_norm": 1.4084968261340196, | |
| "learning_rate": 9.320176335582772e-06, | |
| "loss": 0.3949, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 1.2613636363636362, | |
| "grad_norm": 1.4217514555447899, | |
| "learning_rate": 9.315175493397968e-06, | |
| "loss": 0.466, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 1.2642045454545454, | |
| "grad_norm": 1.378651964217461, | |
| "learning_rate": 9.310157677075847e-06, | |
| "loss": 0.3997, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.2670454545454546, | |
| "grad_norm": 1.4196574298188969, | |
| "learning_rate": 9.30512290635445e-06, | |
| "loss": 0.3745, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 1.2698863636363638, | |
| "grad_norm": 1.3558455660245554, | |
| "learning_rate": 9.300071201038503e-06, | |
| "loss": 0.4164, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 1.2727272727272727, | |
| "grad_norm": 1.3784613104346528, | |
| "learning_rate": 9.295002580999355e-06, | |
| "loss": 0.3385, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 1.2755681818181819, | |
| "grad_norm": 1.4749483610049798, | |
| "learning_rate": 9.289917066174887e-06, | |
| "loss": 0.389, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 1.2784090909090908, | |
| "grad_norm": 1.394529342297257, | |
| "learning_rate": 9.284814676569431e-06, | |
| "loss": 0.4202, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.28125, | |
| "grad_norm": 1.5150361682026248, | |
| "learning_rate": 9.27969543225371e-06, | |
| "loss": 0.4391, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 1.2840909090909092, | |
| "grad_norm": 1.2439757451957247, | |
| "learning_rate": 9.274559353364734e-06, | |
| "loss": 0.3674, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 1.2869318181818181, | |
| "grad_norm": 1.3338341217959293, | |
| "learning_rate": 9.269406460105742e-06, | |
| "loss": 0.4285, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 1.2897727272727273, | |
| "grad_norm": 1.348464897084427, | |
| "learning_rate": 9.264236772746108e-06, | |
| "loss": 0.3749, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 1.2926136363636362, | |
| "grad_norm": 1.532977524300145, | |
| "learning_rate": 9.259050311621274e-06, | |
| "loss": 0.3606, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.2954545454545454, | |
| "grad_norm": 1.2025844811833046, | |
| "learning_rate": 9.253847097132656e-06, | |
| "loss": 0.4191, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 1.2982954545454546, | |
| "grad_norm": 1.2628021212775675, | |
| "learning_rate": 9.248627149747573e-06, | |
| "loss": 0.3696, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 1.3011363636363638, | |
| "grad_norm": 1.424313945946166, | |
| "learning_rate": 9.243390489999166e-06, | |
| "loss": 0.3776, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 1.3039772727272727, | |
| "grad_norm": 1.4193614387826192, | |
| "learning_rate": 9.238137138486318e-06, | |
| "loss": 0.4468, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 1.3068181818181819, | |
| "grad_norm": 1.5337967438868354, | |
| "learning_rate": 9.232867115873566e-06, | |
| "loss": 0.4199, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.3096590909090908, | |
| "grad_norm": 1.4604547826541918, | |
| "learning_rate": 9.227580442891022e-06, | |
| "loss": 0.4003, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 1.3125, | |
| "grad_norm": 1.3483447354558256, | |
| "learning_rate": 9.222277140334301e-06, | |
| "loss": 0.3692, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.3153409090909092, | |
| "grad_norm": 1.3176534368365396, | |
| "learning_rate": 9.21695722906443e-06, | |
| "loss": 0.4082, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.3181818181818181, | |
| "grad_norm": 1.3848734791606088, | |
| "learning_rate": 9.211620730007763e-06, | |
| "loss": 0.4134, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.3210227272727273, | |
| "grad_norm": 1.3812780788636116, | |
| "learning_rate": 9.206267664155906e-06, | |
| "loss": 0.3784, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.3238636363636362, | |
| "grad_norm": 1.2963867485593161, | |
| "learning_rate": 9.200898052565638e-06, | |
| "loss": 0.388, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 1.3267045454545454, | |
| "grad_norm": 1.42585767936649, | |
| "learning_rate": 9.195511916358813e-06, | |
| "loss": 0.3985, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 1.3295454545454546, | |
| "grad_norm": 1.311283985032005, | |
| "learning_rate": 9.19010927672229e-06, | |
| "loss": 0.3719, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 1.3323863636363638, | |
| "grad_norm": 1.297472323865573, | |
| "learning_rate": 9.18469015490785e-06, | |
| "loss": 0.4201, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 1.3352272727272727, | |
| "grad_norm": 1.2429332273904634, | |
| "learning_rate": 9.1792545722321e-06, | |
| "loss": 0.3472, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.3380681818181819, | |
| "grad_norm": 1.3735871193712135, | |
| "learning_rate": 9.173802550076402e-06, | |
| "loss": 0.3834, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 1.3409090909090908, | |
| "grad_norm": 1.1849332008114133, | |
| "learning_rate": 9.168334109886785e-06, | |
| "loss": 0.4078, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 1.34375, | |
| "grad_norm": 1.2435873213180686, | |
| "learning_rate": 9.162849273173857e-06, | |
| "loss": 0.4132, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 1.3465909090909092, | |
| "grad_norm": 1.3880977737093316, | |
| "learning_rate": 9.157348061512728e-06, | |
| "loss": 0.3989, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 1.3494318181818181, | |
| "grad_norm": 1.162608532025314, | |
| "learning_rate": 9.151830496542912e-06, | |
| "loss": 0.3894, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.3522727272727273, | |
| "grad_norm": 1.3820687112561636, | |
| "learning_rate": 9.146296599968258e-06, | |
| "loss": 0.3629, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 1.3551136363636362, | |
| "grad_norm": 1.235383479146175, | |
| "learning_rate": 9.140746393556853e-06, | |
| "loss": 0.3896, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 1.3579545454545454, | |
| "grad_norm": 1.349928335272851, | |
| "learning_rate": 9.135179899140947e-06, | |
| "loss": 0.4773, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.3607954545454546, | |
| "grad_norm": 1.2613773914812836, | |
| "learning_rate": 9.129597138616845e-06, | |
| "loss": 0.3953, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.3636363636363638, | |
| "grad_norm": 1.3516210026791848, | |
| "learning_rate": 9.123998133944854e-06, | |
| "loss": 0.3954, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.3664772727272727, | |
| "grad_norm": 1.4675800730662938, | |
| "learning_rate": 9.118382907149164e-06, | |
| "loss": 0.3922, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.3693181818181819, | |
| "grad_norm": 1.3686097193464257, | |
| "learning_rate": 9.11275148031779e-06, | |
| "loss": 0.4088, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.3721590909090908, | |
| "grad_norm": 1.3124849147092428, | |
| "learning_rate": 9.107103875602458e-06, | |
| "loss": 0.4436, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.375, | |
| "grad_norm": 1.435628812289037, | |
| "learning_rate": 9.101440115218543e-06, | |
| "loss": 0.3658, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.3778409090909092, | |
| "grad_norm": 1.2899778976362926, | |
| "learning_rate": 9.09576022144496e-06, | |
| "loss": 0.4207, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.3806818181818181, | |
| "grad_norm": 1.4371410340359985, | |
| "learning_rate": 9.090064216624093e-06, | |
| "loss": 0.3798, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.3835227272727273, | |
| "grad_norm": 1.2938067062261525, | |
| "learning_rate": 9.084352123161695e-06, | |
| "loss": 0.3966, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.3863636363636362, | |
| "grad_norm": 1.3681448699118202, | |
| "learning_rate": 9.078623963526811e-06, | |
| "loss": 0.4305, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.3892045454545454, | |
| "grad_norm": 1.2589613682277516, | |
| "learning_rate": 9.07287976025168e-06, | |
| "loss": 0.3968, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.3920454545454546, | |
| "grad_norm": 1.3938984533537444, | |
| "learning_rate": 9.067119535931648e-06, | |
| "loss": 0.382, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.3948863636363638, | |
| "grad_norm": 1.2856128305972496, | |
| "learning_rate": 9.061343313225088e-06, | |
| "loss": 0.4349, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.3977272727272727, | |
| "grad_norm": 1.3125963539926566, | |
| "learning_rate": 9.055551114853296e-06, | |
| "loss": 0.44, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.4005681818181819, | |
| "grad_norm": 1.3955500448358946, | |
| "learning_rate": 9.04974296360042e-06, | |
| "loss": 0.4091, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.4034090909090908, | |
| "grad_norm": 1.428413178347213, | |
| "learning_rate": 9.043918882313344e-06, | |
| "loss": 0.4271, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.40625, | |
| "grad_norm": 1.2796124312236525, | |
| "learning_rate": 9.038078893901634e-06, | |
| "loss": 0.4278, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.4090909090909092, | |
| "grad_norm": 1.2349198829939891, | |
| "learning_rate": 9.032223021337415e-06, | |
| "loss": 0.3759, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.4119318181818181, | |
| "grad_norm": 1.3595968193610075, | |
| "learning_rate": 9.026351287655294e-06, | |
| "loss": 0.4327, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.4147727272727273, | |
| "grad_norm": 1.5092161593843314, | |
| "learning_rate": 9.020463715952276e-06, | |
| "loss": 0.4251, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.4176136363636362, | |
| "grad_norm": 1.2943083527048878, | |
| "learning_rate": 9.014560329387661e-06, | |
| "loss": 0.4012, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.4204545454545454, | |
| "grad_norm": 1.3181954379592955, | |
| "learning_rate": 9.00864115118296e-06, | |
| "loss": 0.4113, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.4232954545454546, | |
| "grad_norm": 1.2580904992249191, | |
| "learning_rate": 9.002706204621802e-06, | |
| "loss": 0.4157, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.4261363636363638, | |
| "grad_norm": 1.352479946463779, | |
| "learning_rate": 8.996755513049844e-06, | |
| "loss": 0.4393, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.4289772727272727, | |
| "grad_norm": 1.3576556234541397, | |
| "learning_rate": 8.99078909987467e-06, | |
| "loss": 0.3755, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.4318181818181819, | |
| "grad_norm": 1.230294030046518, | |
| "learning_rate": 8.984806988565716e-06, | |
| "loss": 0.3966, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.4346590909090908, | |
| "grad_norm": 1.272378129867815, | |
| "learning_rate": 8.978809202654161e-06, | |
| "loss": 0.3982, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.4375, | |
| "grad_norm": 1.4640507964444314, | |
| "learning_rate": 8.972795765732847e-06, | |
| "loss": 0.4077, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.4403409090909092, | |
| "grad_norm": 1.4122007567764803, | |
| "learning_rate": 8.966766701456177e-06, | |
| "loss": 0.376, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.4431818181818181, | |
| "grad_norm": 1.2543181499121292, | |
| "learning_rate": 8.960722033540025e-06, | |
| "loss": 0.3524, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.4460227272727273, | |
| "grad_norm": 1.3516761909069253, | |
| "learning_rate": 8.954661785761648e-06, | |
| "loss": 0.3809, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.4488636363636362, | |
| "grad_norm": 1.3738331770015468, | |
| "learning_rate": 8.94858598195958e-06, | |
| "loss": 0.4381, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.4517045454545454, | |
| "grad_norm": 1.3894759101552216, | |
| "learning_rate": 8.942494646033555e-06, | |
| "loss": 0.4007, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.4545454545454546, | |
| "grad_norm": 1.3813101187524204, | |
| "learning_rate": 8.936387801944397e-06, | |
| "loss": 0.4241, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.4573863636363638, | |
| "grad_norm": 1.234475792865503, | |
| "learning_rate": 8.930265473713939e-06, | |
| "loss": 0.3475, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.4602272727272727, | |
| "grad_norm": 1.340171681220909, | |
| "learning_rate": 8.924127685424914e-06, | |
| "loss": 0.3837, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.4630681818181819, | |
| "grad_norm": 1.611032851097366, | |
| "learning_rate": 8.917974461220877e-06, | |
| "loss": 0.4447, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.4659090909090908, | |
| "grad_norm": 1.4903364424496537, | |
| "learning_rate": 8.911805825306097e-06, | |
| "loss": 0.4018, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.46875, | |
| "grad_norm": 1.3377419588076225, | |
| "learning_rate": 8.905621801945467e-06, | |
| "loss": 0.3697, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.4715909090909092, | |
| "grad_norm": 1.2820938683900953, | |
| "learning_rate": 8.899422415464409e-06, | |
| "loss": 0.4119, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.4744318181818181, | |
| "grad_norm": 1.1786452967753045, | |
| "learning_rate": 8.893207690248776e-06, | |
| "loss": 0.4287, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.4772727272727273, | |
| "grad_norm": 1.3659687009025128, | |
| "learning_rate": 8.88697765074476e-06, | |
| "loss": 0.4042, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.4801136363636362, | |
| "grad_norm": 1.4456826342268947, | |
| "learning_rate": 8.880732321458785e-06, | |
| "loss": 0.4582, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.4829545454545454, | |
| "grad_norm": 1.3518179145479006, | |
| "learning_rate": 8.87447172695743e-06, | |
| "loss": 0.4094, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.4857954545454546, | |
| "grad_norm": 1.32516388869156, | |
| "learning_rate": 8.868195891867315e-06, | |
| "loss": 0.421, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.4886363636363638, | |
| "grad_norm": 1.4216865156200345, | |
| "learning_rate": 8.86190484087501e-06, | |
| "loss": 0.4337, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.4914772727272727, | |
| "grad_norm": 1.2427141581558576, | |
| "learning_rate": 8.85559859872694e-06, | |
| "loss": 0.3898, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.4943181818181819, | |
| "grad_norm": 1.3595678928667188, | |
| "learning_rate": 8.849277190229284e-06, | |
| "loss": 0.3784, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.4971590909090908, | |
| "grad_norm": 1.3554942914251902, | |
| "learning_rate": 8.84294064024788e-06, | |
| "loss": 0.4704, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 1.2506191801860325, | |
| "learning_rate": 8.836588973708129e-06, | |
| "loss": 0.4225, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.5028409090909092, | |
| "grad_norm": 1.272132348825335, | |
| "learning_rate": 8.83022221559489e-06, | |
| "loss": 0.4212, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.5056818181818183, | |
| "grad_norm": 1.2555892277001355, | |
| "learning_rate": 8.82384039095239e-06, | |
| "loss": 0.4531, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.5085227272727273, | |
| "grad_norm": 1.4108423232568015, | |
| "learning_rate": 8.817443524884119e-06, | |
| "loss": 0.3875, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.5113636363636362, | |
| "grad_norm": 1.5126815712719472, | |
| "learning_rate": 8.811031642552732e-06, | |
| "loss": 0.3704, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.5142045454545454, | |
| "grad_norm": 1.4219068937968729, | |
| "learning_rate": 8.804604769179958e-06, | |
| "loss": 0.3904, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.5170454545454546, | |
| "grad_norm": 1.3044529230830835, | |
| "learning_rate": 8.798162930046488e-06, | |
| "loss": 0.3762, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.5198863636363638, | |
| "grad_norm": 1.4796842484308415, | |
| "learning_rate": 8.791706150491887e-06, | |
| "loss": 0.3911, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.5227272727272727, | |
| "grad_norm": 1.4946259268999356, | |
| "learning_rate": 8.78523445591449e-06, | |
| "loss": 0.3734, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.5255681818181817, | |
| "grad_norm": 1.5186972153108664, | |
| "learning_rate": 8.778747871771293e-06, | |
| "loss": 0.419, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.5284090909090908, | |
| "grad_norm": 1.2896009388954741, | |
| "learning_rate": 8.772246423577871e-06, | |
| "loss": 0.435, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.53125, | |
| "grad_norm": 1.4805755697726184, | |
| "learning_rate": 8.765730136908266e-06, | |
| "loss": 0.3916, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.5340909090909092, | |
| "grad_norm": 1.4854240889161028, | |
| "learning_rate": 8.759199037394888e-06, | |
| "loss": 0.3848, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.5369318181818183, | |
| "grad_norm": 1.463681525472631, | |
| "learning_rate": 8.752653150728412e-06, | |
| "loss": 0.4095, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.5397727272727273, | |
| "grad_norm": 1.5820724211894255, | |
| "learning_rate": 8.746092502657681e-06, | |
| "loss": 0.4613, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.5426136363636362, | |
| "grad_norm": 1.505592006271489, | |
| "learning_rate": 8.739517118989606e-06, | |
| "loss": 0.3683, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.5454545454545454, | |
| "grad_norm": 1.3692299981481657, | |
| "learning_rate": 8.732927025589058e-06, | |
| "loss": 0.3586, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.5482954545454546, | |
| "grad_norm": 1.3878049908700243, | |
| "learning_rate": 8.726322248378775e-06, | |
| "loss": 0.3788, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.5511363636363638, | |
| "grad_norm": 1.3928801140937896, | |
| "learning_rate": 8.719702813339248e-06, | |
| "loss": 0.4042, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.5539772727272727, | |
| "grad_norm": 1.2922560147345834, | |
| "learning_rate": 8.713068746508633e-06, | |
| "loss": 0.4193, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.5568181818181817, | |
| "grad_norm": 1.4892914844476706, | |
| "learning_rate": 8.706420073982636e-06, | |
| "loss": 0.4617, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.5596590909090908, | |
| "grad_norm": 1.4717274089795425, | |
| "learning_rate": 8.69975682191442e-06, | |
| "loss": 0.3646, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.5625, | |
| "grad_norm": 1.4582367131460128, | |
| "learning_rate": 8.693079016514497e-06, | |
| "loss": 0.4224, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.5653409090909092, | |
| "grad_norm": 1.3700688358872546, | |
| "learning_rate": 8.68638668405062e-06, | |
| "loss": 0.3761, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.5681818181818183, | |
| "grad_norm": 1.303150019049142, | |
| "learning_rate": 8.679679850847698e-06, | |
| "loss": 0.3941, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.5710227272727273, | |
| "grad_norm": 1.3691241123764724, | |
| "learning_rate": 8.672958543287666e-06, | |
| "loss": 0.4003, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.5738636363636362, | |
| "grad_norm": 1.3220887396048668, | |
| "learning_rate": 8.666222787809407e-06, | |
| "loss": 0.4369, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.5767045454545454, | |
| "grad_norm": 1.3926991655468106, | |
| "learning_rate": 8.659472610908628e-06, | |
| "loss": 0.3942, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.5795454545454546, | |
| "grad_norm": 1.3537640175865195, | |
| "learning_rate": 8.652708039137767e-06, | |
| "loss": 0.3829, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.5823863636363638, | |
| "grad_norm": 1.3430604939036295, | |
| "learning_rate": 8.645929099105886e-06, | |
| "loss": 0.4007, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.5852272727272727, | |
| "grad_norm": 1.236166779871983, | |
| "learning_rate": 8.639135817478566e-06, | |
| "loss": 0.3855, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.5880681818181817, | |
| "grad_norm": 1.2858592765808123, | |
| "learning_rate": 8.632328220977801e-06, | |
| "loss": 0.3809, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.5909090909090908, | |
| "grad_norm": 1.346749375703671, | |
| "learning_rate": 8.625506336381895e-06, | |
| "loss": 0.3927, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.59375, | |
| "grad_norm": 1.3865417442370538, | |
| "learning_rate": 8.61867019052535e-06, | |
| "loss": 0.406, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.5965909090909092, | |
| "grad_norm": 1.3928703341415491, | |
| "learning_rate": 8.611819810298778e-06, | |
| "loss": 0.4348, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.5994318181818183, | |
| "grad_norm": 1.3346447546562574, | |
| "learning_rate": 8.604955222648772e-06, | |
| "loss": 0.3923, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.6022727272727273, | |
| "grad_norm": 1.3874781514983023, | |
| "learning_rate": 8.598076454577815e-06, | |
| "loss": 0.3811, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.6051136363636362, | |
| "grad_norm": 1.3892030370722708, | |
| "learning_rate": 8.591183533144172e-06, | |
| "loss": 0.4168, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.6079545454545454, | |
| "grad_norm": 1.39476614799105, | |
| "learning_rate": 8.584276485461775e-06, | |
| "loss": 0.3924, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.6107954545454546, | |
| "grad_norm": 1.4734200006566598, | |
| "learning_rate": 8.577355338700133e-06, | |
| "loss": 0.3808, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.6136363636363638, | |
| "grad_norm": 1.4118837014643815, | |
| "learning_rate": 8.570420120084208e-06, | |
| "loss": 0.456, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.6164772727272727, | |
| "grad_norm": 1.3035533880676229, | |
| "learning_rate": 8.563470856894316e-06, | |
| "loss": 0.4052, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.6193181818181817, | |
| "grad_norm": 1.2585997343091353, | |
| "learning_rate": 8.556507576466017e-06, | |
| "loss": 0.4055, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.6221590909090908, | |
| "grad_norm": 1.2827553058465992, | |
| "learning_rate": 8.549530306190015e-06, | |
| "loss": 0.4242, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.625, | |
| "grad_norm": 1.311108635707505, | |
| "learning_rate": 8.542539073512038e-06, | |
| "loss": 0.4325, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.6278409090909092, | |
| "grad_norm": 1.452722602140385, | |
| "learning_rate": 8.535533905932739e-06, | |
| "loss": 0.359, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.6306818181818183, | |
| "grad_norm": 1.4216842513339472, | |
| "learning_rate": 8.528514831007587e-06, | |
| "loss": 0.3865, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.6335227272727273, | |
| "grad_norm": 1.104281881309897, | |
| "learning_rate": 8.521481876346751e-06, | |
| "loss": 0.3927, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.6363636363636362, | |
| "grad_norm": 1.373540934205005, | |
| "learning_rate": 8.514435069615005e-06, | |
| "loss": 0.3735, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.6392045454545454, | |
| "grad_norm": 1.4721648721287588, | |
| "learning_rate": 8.507374438531606e-06, | |
| "loss": 0.4089, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.6420454545454546, | |
| "grad_norm": 1.4330241380765012, | |
| "learning_rate": 8.500300010870195e-06, | |
| "loss": 0.3962, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.6448863636363638, | |
| "grad_norm": 1.314258656388838, | |
| "learning_rate": 8.493211814458674e-06, | |
| "loss": 0.3902, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.6477272727272727, | |
| "grad_norm": 1.3180359191061783, | |
| "learning_rate": 8.486109877179115e-06, | |
| "loss": 0.4559, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.6505681818181817, | |
| "grad_norm": 1.2083828163470816, | |
| "learning_rate": 8.478994226967638e-06, | |
| "loss": 0.3928, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.6534090909090908, | |
| "grad_norm": 1.274567661010079, | |
| "learning_rate": 8.471864891814304e-06, | |
| "loss": 0.4143, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.65625, | |
| "grad_norm": 1.3144626541356723, | |
| "learning_rate": 8.464721899763003e-06, | |
| "loss": 0.3369, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.6590909090909092, | |
| "grad_norm": 1.3987801501809722, | |
| "learning_rate": 8.457565278911349e-06, | |
| "loss": 0.4459, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.6619318181818183, | |
| "grad_norm": 1.235004230998312, | |
| "learning_rate": 8.450395057410561e-06, | |
| "loss": 0.3851, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.6647727272727273, | |
| "grad_norm": 1.3390522796730047, | |
| "learning_rate": 8.443211263465362e-06, | |
| "loss": 0.3945, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.6676136363636362, | |
| "grad_norm": 1.4255845992278726, | |
| "learning_rate": 8.436013925333868e-06, | |
| "loss": 0.4216, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.6704545454545454, | |
| "grad_norm": 1.382760628657945, | |
| "learning_rate": 8.42880307132746e-06, | |
| "loss": 0.3993, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.6732954545454546, | |
| "grad_norm": 1.3285153601515391, | |
| "learning_rate": 8.421578729810693e-06, | |
| "loss": 0.4177, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.6761363636363638, | |
| "grad_norm": 1.2999965518088938, | |
| "learning_rate": 8.414340929201175e-06, | |
| "loss": 0.424, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.6789772727272727, | |
| "grad_norm": 1.2296769461458794, | |
| "learning_rate": 8.407089697969458e-06, | |
| "loss": 0.4055, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.6818181818181817, | |
| "grad_norm": 1.249775689774472, | |
| "learning_rate": 8.39982506463892e-06, | |
| "loss": 0.3783, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.6846590909090908, | |
| "grad_norm": 1.3164118726921976, | |
| "learning_rate": 8.392547057785662e-06, | |
| "loss": 0.4116, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.6875, | |
| "grad_norm": 1.3465959411917943, | |
| "learning_rate": 8.38525570603839e-06, | |
| "loss": 0.4228, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.6903409090909092, | |
| "grad_norm": 1.3714243038000036, | |
| "learning_rate": 8.377951038078303e-06, | |
| "loss": 0.4441, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.6931818181818183, | |
| "grad_norm": 1.334863296659028, | |
| "learning_rate": 8.370633082638977e-06, | |
| "loss": 0.3498, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.6960227272727273, | |
| "grad_norm": 1.3740648201371748, | |
| "learning_rate": 8.363301868506264e-06, | |
| "loss": 0.4077, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.6988636363636362, | |
| "grad_norm": 1.2784134213122538, | |
| "learning_rate": 8.35595742451816e-06, | |
| "loss": 0.412, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.7017045454545454, | |
| "grad_norm": 1.2139020970296652, | |
| "learning_rate": 8.34859977956471e-06, | |
| "loss": 0.4094, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.7045454545454546, | |
| "grad_norm": 1.3204896950697897, | |
| "learning_rate": 8.341228962587881e-06, | |
| "loss": 0.3752, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.7073863636363638, | |
| "grad_norm": 1.4708757740664673, | |
| "learning_rate": 8.33384500258146e-06, | |
| "loss": 0.389, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.7102272727272727, | |
| "grad_norm": 1.4266850812841863, | |
| "learning_rate": 8.326447928590921e-06, | |
| "loss": 0.4396, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.7130681818181817, | |
| "grad_norm": 1.1175427095357402, | |
| "learning_rate": 8.319037769713338e-06, | |
| "loss": 0.382, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.7159090909090908, | |
| "grad_norm": 1.281173570535696, | |
| "learning_rate": 8.311614555097245e-06, | |
| "loss": 0.4192, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.71875, | |
| "grad_norm": 1.240781935793036, | |
| "learning_rate": 8.304178313942536e-06, | |
| "loss": 0.3941, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.7215909090909092, | |
| "grad_norm": 1.3452354473216068, | |
| "learning_rate": 8.296729075500345e-06, | |
| "loss": 0.4161, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.7244318181818183, | |
| "grad_norm": 1.3547326876440298, | |
| "learning_rate": 8.289266869072933e-06, | |
| "loss": 0.4318, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.7272727272727273, | |
| "grad_norm": 1.3778793045074398, | |
| "learning_rate": 8.281791724013571e-06, | |
| "loss": 0.345, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.7301136363636362, | |
| "grad_norm": 1.4145733189261174, | |
| "learning_rate": 8.274303669726427e-06, | |
| "loss": 0.4236, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.7329545454545454, | |
| "grad_norm": 1.4088839074475497, | |
| "learning_rate": 8.266802735666443e-06, | |
| "loss": 0.4407, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.7357954545454546, | |
| "grad_norm": 1.3119328043721623, | |
| "learning_rate": 8.259288951339233e-06, | |
| "loss": 0.3939, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.7386363636363638, | |
| "grad_norm": 1.309429912330872, | |
| "learning_rate": 8.251762346300954e-06, | |
| "loss": 0.4109, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.7414772727272727, | |
| "grad_norm": 1.3345438738994535, | |
| "learning_rate": 8.244222950158194e-06, | |
| "loss": 0.4536, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.7443181818181817, | |
| "grad_norm": 1.453082925840765, | |
| "learning_rate": 8.236670792567856e-06, | |
| "loss": 0.4107, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.7471590909090908, | |
| "grad_norm": 1.32130004870712, | |
| "learning_rate": 8.229105903237045e-06, | |
| "loss": 0.3599, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "grad_norm": 1.2225515703742833, | |
| "learning_rate": 8.221528311922941e-06, | |
| "loss": 0.4239, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.7528409090909092, | |
| "grad_norm": 1.250676696637379, | |
| "learning_rate": 8.213938048432697e-06, | |
| "loss": 0.378, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.7556818181818183, | |
| "grad_norm": 1.3753326745883003, | |
| "learning_rate": 8.206335142623305e-06, | |
| "loss": 0.4368, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.7585227272727273, | |
| "grad_norm": 1.4771926734091496, | |
| "learning_rate": 8.198719624401493e-06, | |
| "loss": 0.408, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.7613636363636362, | |
| "grad_norm": 1.3052916853608703, | |
| "learning_rate": 8.191091523723594e-06, | |
| "loss": 0.3869, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.7642045454545454, | |
| "grad_norm": 1.2079205005036262, | |
| "learning_rate": 8.183450870595443e-06, | |
| "loss": 0.3799, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.7670454545454546, | |
| "grad_norm": 1.3981335006171198, | |
| "learning_rate": 8.175797695072245e-06, | |
| "loss": 0.4104, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.7698863636363638, | |
| "grad_norm": 1.2549497514899004, | |
| "learning_rate": 8.168132027258467e-06, | |
| "loss": 0.3949, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.7727272727272727, | |
| "grad_norm": 1.1515091844850416, | |
| "learning_rate": 8.160453897307714e-06, | |
| "loss": 0.4306, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.7755681818181817, | |
| "grad_norm": 1.260271278484822, | |
| "learning_rate": 8.152763335422612e-06, | |
| "loss": 0.4018, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.7784090909090908, | |
| "grad_norm": 1.2456788115801498, | |
| "learning_rate": 8.145060371854692e-06, | |
| "loss": 0.4038, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.78125, | |
| "grad_norm": 1.3353759651701624, | |
| "learning_rate": 8.13734503690426e-06, | |
| "loss": 0.4053, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.7840909090909092, | |
| "grad_norm": 1.2141575196798886, | |
| "learning_rate": 8.129617360920297e-06, | |
| "loss": 0.374, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.7869318181818183, | |
| "grad_norm": 1.2568084197585532, | |
| "learning_rate": 8.121877374300318e-06, | |
| "loss": 0.3789, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.7897727272727273, | |
| "grad_norm": 1.3228386855112886, | |
| "learning_rate": 8.11412510749027e-06, | |
| "loss": 0.4189, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.7926136363636362, | |
| "grad_norm": 1.2450848434801394, | |
| "learning_rate": 8.106360590984406e-06, | |
| "loss": 0.4058, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.7954545454545454, | |
| "grad_norm": 1.2751134284940169, | |
| "learning_rate": 8.098583855325157e-06, | |
| "loss": 0.3609, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.7982954545454546, | |
| "grad_norm": 1.2613593262795375, | |
| "learning_rate": 8.090794931103026e-06, | |
| "loss": 0.4208, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.8011363636363638, | |
| "grad_norm": 1.3187892972599105, | |
| "learning_rate": 8.08299384895646e-06, | |
| "loss": 0.3644, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.8039772727272727, | |
| "grad_norm": 1.3412404440839794, | |
| "learning_rate": 8.075180639571726e-06, | |
| "loss": 0.3943, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.8068181818181817, | |
| "grad_norm": 1.3416612818936742, | |
| "learning_rate": 8.067355333682799e-06, | |
| "loss": 0.3722, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.8096590909090908, | |
| "grad_norm": 1.2215225081773167, | |
| "learning_rate": 8.059517962071234e-06, | |
| "loss": 0.373, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.8125, | |
| "grad_norm": 1.2968614196474024, | |
| "learning_rate": 8.05166855556605e-06, | |
| "loss": 0.3954, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.8153409090909092, | |
| "grad_norm": 1.2381952159486795, | |
| "learning_rate": 8.043807145043604e-06, | |
| "loss": 0.4333, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 1.3448593265426114, | |
| "learning_rate": 8.035933761427475e-06, | |
| "loss": 0.3933, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.8210227272727273, | |
| "grad_norm": 1.3928700734438324, | |
| "learning_rate": 8.028048435688333e-06, | |
| "loss": 0.3611, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.8238636363636362, | |
| "grad_norm": 1.324880463233581, | |
| "learning_rate": 8.020151198843833e-06, | |
| "loss": 0.3882, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.8267045454545454, | |
| "grad_norm": 1.2671162081281926, | |
| "learning_rate": 8.012242081958477e-06, | |
| "loss": 0.3475, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.8295454545454546, | |
| "grad_norm": 1.3107781953106952, | |
| "learning_rate": 8.004321116143496e-06, | |
| "loss": 0.3596, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.8323863636363638, | |
| "grad_norm": 1.3666505259241546, | |
| "learning_rate": 7.996388332556735e-06, | |
| "loss": 0.4169, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.8352272727272727, | |
| "grad_norm": 1.4514440975950802, | |
| "learning_rate": 7.988443762402525e-06, | |
| "loss": 0.3918, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.8380681818181817, | |
| "grad_norm": 1.2979859649109158, | |
| "learning_rate": 7.980487436931558e-06, | |
| "loss": 0.4341, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.8409090909090908, | |
| "grad_norm": 1.2503565961960492, | |
| "learning_rate": 7.972519387440767e-06, | |
| "loss": 0.4058, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.84375, | |
| "grad_norm": 1.2538862467599425, | |
| "learning_rate": 7.964539645273204e-06, | |
| "loss": 0.4106, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.8465909090909092, | |
| "grad_norm": 1.2837654267836336, | |
| "learning_rate": 7.956548241817914e-06, | |
| "loss": 0.3882, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.8494318181818183, | |
| "grad_norm": 1.283512404811162, | |
| "learning_rate": 7.948545208509811e-06, | |
| "loss": 0.4181, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.8522727272727273, | |
| "grad_norm": 1.3386923463737823, | |
| "learning_rate": 7.940530576829562e-06, | |
| "loss": 0.3923, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.8551136363636362, | |
| "grad_norm": 1.2341509248969929, | |
| "learning_rate": 7.932504378303452e-06, | |
| "loss": 0.365, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.8579545454545454, | |
| "grad_norm": 1.216460582958605, | |
| "learning_rate": 7.924466644503265e-06, | |
| "loss": 0.3495, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.8607954545454546, | |
| "grad_norm": 1.3130118175509913, | |
| "learning_rate": 7.916417407046166e-06, | |
| "loss": 0.3891, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.8636363636363638, | |
| "grad_norm": 1.3009995160105379, | |
| "learning_rate": 7.908356697594562e-06, | |
| "loss": 0.3555, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.8664772727272727, | |
| "grad_norm": 1.2983355016455334, | |
| "learning_rate": 7.900284547855992e-06, | |
| "loss": 0.4183, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.8693181818181817, | |
| "grad_norm": 1.3025519454264027, | |
| "learning_rate": 7.892200989582994e-06, | |
| "loss": 0.3832, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.8721590909090908, | |
| "grad_norm": 1.4015896773460164, | |
| "learning_rate": 7.884106054572987e-06, | |
| "loss": 0.3852, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 1.3658709627701167, | |
| "learning_rate": 7.875999774668135e-06, | |
| "loss": 0.4103, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.8778409090909092, | |
| "grad_norm": 1.1749326223713263, | |
| "learning_rate": 7.86788218175523e-06, | |
| "loss": 0.4185, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.8806818181818183, | |
| "grad_norm": 1.3300104427508925, | |
| "learning_rate": 7.859753307765571e-06, | |
| "loss": 0.4146, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.8835227272727273, | |
| "grad_norm": 1.3350244520568906, | |
| "learning_rate": 7.851613184674821e-06, | |
| "loss": 0.3795, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.8863636363636362, | |
| "grad_norm": 1.3114376934881007, | |
| "learning_rate": 7.843461844502903e-06, | |
| "loss": 0.3887, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.8892045454545454, | |
| "grad_norm": 1.263848611253203, | |
| "learning_rate": 7.835299319313854e-06, | |
| "loss": 0.3627, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.8920454545454546, | |
| "grad_norm": 1.3067022373616142, | |
| "learning_rate": 7.827125641215718e-06, | |
| "loss": 0.4148, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.8948863636363638, | |
| "grad_norm": 1.3313015879998442, | |
| "learning_rate": 7.818940842360404e-06, | |
| "loss": 0.4221, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.8977272727272727, | |
| "grad_norm": 1.286432747246891, | |
| "learning_rate": 7.810744954943564e-06, | |
| "loss": 0.3851, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.9005681818181817, | |
| "grad_norm": 1.2339003550511074, | |
| "learning_rate": 7.80253801120447e-06, | |
| "loss": 0.3771, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.9034090909090908, | |
| "grad_norm": 1.2667417343121634, | |
| "learning_rate": 7.79432004342589e-06, | |
| "loss": 0.4144, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.90625, | |
| "grad_norm": 1.3970266065185077, | |
| "learning_rate": 7.78609108393395e-06, | |
| "loss": 0.3739, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.9090909090909092, | |
| "grad_norm": 1.40544283943565, | |
| "learning_rate": 7.777851165098012e-06, | |
| "loss": 0.446, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.9119318181818183, | |
| "grad_norm": 1.3511713634953924, | |
| "learning_rate": 7.769600319330553e-06, | |
| "loss": 0.4265, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.9147727272727273, | |
| "grad_norm": 1.362587721165509, | |
| "learning_rate": 7.761338579087026e-06, | |
| "loss": 0.4179, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.9176136363636362, | |
| "grad_norm": 1.3317227014945128, | |
| "learning_rate": 7.753065976865745e-06, | |
| "loss": 0.444, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.9204545454545454, | |
| "grad_norm": 1.353251061657008, | |
| "learning_rate": 7.744782545207745e-06, | |
| "loss": 0.4288, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.9232954545454546, | |
| "grad_norm": 1.3310135379731283, | |
| "learning_rate": 7.736488316696663e-06, | |
| "loss": 0.363, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.9261363636363638, | |
| "grad_norm": 1.4336597059726608, | |
| "learning_rate": 7.728183323958603e-06, | |
| "loss": 0.3774, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.9289772727272727, | |
| "grad_norm": 1.3096730175462812, | |
| "learning_rate": 7.719867599662017e-06, | |
| "loss": 0.3815, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.9318181818181817, | |
| "grad_norm": 1.2452300162277348, | |
| "learning_rate": 7.711541176517563e-06, | |
| "loss": 0.3523, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.9346590909090908, | |
| "grad_norm": 1.2601600689651646, | |
| "learning_rate": 7.703204087277989e-06, | |
| "loss": 0.4133, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.9375, | |
| "grad_norm": 1.3131138015057662, | |
| "learning_rate": 7.694856364737997e-06, | |
| "loss": 0.3528, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.9403409090909092, | |
| "grad_norm": 1.3174274202800598, | |
| "learning_rate": 7.686498041734121e-06, | |
| "loss": 0.3687, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.9431818181818183, | |
| "grad_norm": 1.382354479301822, | |
| "learning_rate": 7.678129151144582e-06, | |
| "loss": 0.385, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.9460227272727273, | |
| "grad_norm": 1.342621819895786, | |
| "learning_rate": 7.669749725889182e-06, | |
| "loss": 0.4192, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.9488636363636362, | |
| "grad_norm": 1.2633452874742823, | |
| "learning_rate": 7.661359798929152e-06, | |
| "loss": 0.4359, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.9517045454545454, | |
| "grad_norm": 1.4034254507755521, | |
| "learning_rate": 7.65295940326704e-06, | |
| "loss": 0.4036, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.9545454545454546, | |
| "grad_norm": 1.2499008102214813, | |
| "learning_rate": 7.644548571946569e-06, | |
| "loss": 0.402, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.9573863636363638, | |
| "grad_norm": 1.4019764959578607, | |
| "learning_rate": 7.636127338052513e-06, | |
| "loss": 0.4168, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.9602272727272727, | |
| "grad_norm": 1.3111166839375912, | |
| "learning_rate": 7.627695734710565e-06, | |
| "loss": 0.4128, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.9630681818181817, | |
| "grad_norm": 1.2665556025054903, | |
| "learning_rate": 7.619253795087209e-06, | |
| "loss": 0.3456, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.9659090909090908, | |
| "grad_norm": 1.221754638519089, | |
| "learning_rate": 7.610801552389584e-06, | |
| "loss": 0.4104, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.96875, | |
| "grad_norm": 1.1457951321520503, | |
| "learning_rate": 7.602339039865362e-06, | |
| "loss": 0.3633, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.9715909090909092, | |
| "grad_norm": 1.344478912554315, | |
| "learning_rate": 7.593866290802608e-06, | |
| "loss": 0.451, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.9744318181818183, | |
| "grad_norm": 1.4122647415470586, | |
| "learning_rate": 7.5853833385296545e-06, | |
| "loss": 0.3944, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.9772727272727273, | |
| "grad_norm": 1.4648734442255729, | |
| "learning_rate": 7.576890216414973e-06, | |
| "loss": 0.4243, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.9801136363636362, | |
| "grad_norm": 1.1880578823116779, | |
| "learning_rate": 7.568386957867033e-06, | |
| "loss": 0.3682, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.9829545454545454, | |
| "grad_norm": 1.388741245940409, | |
| "learning_rate": 7.559873596334179e-06, | |
| "loss": 0.3952, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.9857954545454546, | |
| "grad_norm": 1.3426340364215053, | |
| "learning_rate": 7.5513501653045e-06, | |
| "loss": 0.3859, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.9886363636363638, | |
| "grad_norm": 1.342227819106623, | |
| "learning_rate": 7.542816698305686e-06, | |
| "loss": 0.3573, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.9914772727272727, | |
| "grad_norm": 1.2713457352199766, | |
| "learning_rate": 7.534273228904916e-06, | |
| "loss": 0.3714, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.9943181818181817, | |
| "grad_norm": 1.2814840874973006, | |
| "learning_rate": 7.525719790708703e-06, | |
| "loss": 0.4265, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.9971590909090908, | |
| "grad_norm": 1.224475509495077, | |
| "learning_rate": 7.5171564173627795e-06, | |
| "loss": 0.3623, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 1.3940157665360537, | |
| "learning_rate": 7.508583142551959e-06, | |
| "loss": 0.3359, | |
| "step": 704 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1760, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 352, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 727389995073536.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |