| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.9943851768669285, |
| "eval_steps": 500, |
| "global_step": 666, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.004491858506457047, |
| "grad_norm": 5.747499942779541, |
| "learning_rate": 1.4925373134328358e-07, |
| "loss": 0.8287, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.008983717012914094, |
| "grad_norm": 5.93737268447876, |
| "learning_rate": 2.9850746268656716e-07, |
| "loss": 0.8732, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.01347557551937114, |
| "grad_norm": 5.791993618011475, |
| "learning_rate": 4.4776119402985074e-07, |
| "loss": 0.8552, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.017967434025828188, |
| "grad_norm": 5.838237762451172, |
| "learning_rate": 5.970149253731343e-07, |
| "loss": 0.8678, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.022459292532285232, |
| "grad_norm": 5.8886027336120605, |
| "learning_rate": 7.462686567164179e-07, |
| "loss": 0.8734, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.02695115103874228, |
| "grad_norm": 5.727405071258545, |
| "learning_rate": 8.955223880597015e-07, |
| "loss": 0.8517, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.031443009545199324, |
| "grad_norm": 5.336740493774414, |
| "learning_rate": 1.044776119402985e-06, |
| "loss": 0.8407, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.035934868051656375, |
| "grad_norm": 5.36053466796875, |
| "learning_rate": 1.1940298507462686e-06, |
| "loss": 0.8449, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.04042672655811342, |
| "grad_norm": 4.371918678283691, |
| "learning_rate": 1.3432835820895524e-06, |
| "loss": 0.8134, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.044918585064570464, |
| "grad_norm": 4.224687099456787, |
| "learning_rate": 1.4925373134328358e-06, |
| "loss": 0.809, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.049410443571027515, |
| "grad_norm": 3.9456374645233154, |
| "learning_rate": 1.6417910447761196e-06, |
| "loss": 0.8054, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.05390230207748456, |
| "grad_norm": 2.2551746368408203, |
| "learning_rate": 1.791044776119403e-06, |
| "loss": 0.7676, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.058394160583941604, |
| "grad_norm": 2.1659226417541504, |
| "learning_rate": 1.9402985074626867e-06, |
| "loss": 0.7585, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.06288601909039865, |
| "grad_norm": 1.9352374076843262, |
| "learning_rate": 2.08955223880597e-06, |
| "loss": 0.759, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.06737787759685569, |
| "grad_norm": 1.8645031452178955, |
| "learning_rate": 2.238805970149254e-06, |
| "loss": 0.7515, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.07186973610331275, |
| "grad_norm": 2.0890092849731445, |
| "learning_rate": 2.3880597014925373e-06, |
| "loss": 0.7335, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.0763615946097698, |
| "grad_norm": 3.110126256942749, |
| "learning_rate": 2.537313432835821e-06, |
| "loss": 0.7421, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.08085345311622684, |
| "grad_norm": 3.361837148666382, |
| "learning_rate": 2.686567164179105e-06, |
| "loss": 0.7311, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.08534531162268388, |
| "grad_norm": 3.211651086807251, |
| "learning_rate": 2.835820895522388e-06, |
| "loss": 0.7229, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.08983717012914093, |
| "grad_norm": 3.045382022857666, |
| "learning_rate": 2.9850746268656716e-06, |
| "loss": 0.7301, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.09432902863559797, |
| "grad_norm": 2.593707799911499, |
| "learning_rate": 3.1343283582089558e-06, |
| "loss": 0.7085, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.09882088714205503, |
| "grad_norm": 1.6424776315689087, |
| "learning_rate": 3.283582089552239e-06, |
| "loss": 0.6902, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.10331274564851207, |
| "grad_norm": 1.2647095918655396, |
| "learning_rate": 3.4328358208955225e-06, |
| "loss": 0.6807, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.10780460415496912, |
| "grad_norm": 1.1175804138183594, |
| "learning_rate": 3.582089552238806e-06, |
| "loss": 0.6662, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.11229646266142616, |
| "grad_norm": 1.0176746845245361, |
| "learning_rate": 3.73134328358209e-06, |
| "loss": 0.6643, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.11678832116788321, |
| "grad_norm": 0.9287970066070557, |
| "learning_rate": 3.8805970149253735e-06, |
| "loss": 0.6396, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.12128017967434025, |
| "grad_norm": 1.0267021656036377, |
| "learning_rate": 4.029850746268657e-06, |
| "loss": 0.6461, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.1257720381807973, |
| "grad_norm": 0.9184905886650085, |
| "learning_rate": 4.17910447761194e-06, |
| "loss": 0.6439, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.13026389668725435, |
| "grad_norm": 0.7492115497589111, |
| "learning_rate": 4.3283582089552236e-06, |
| "loss": 0.6403, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.13475575519371139, |
| "grad_norm": 0.6593955159187317, |
| "learning_rate": 4.477611940298508e-06, |
| "loss": 0.6358, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.13924761370016844, |
| "grad_norm": 0.6903685927391052, |
| "learning_rate": 4.626865671641791e-06, |
| "loss": 0.6318, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.1437394722066255, |
| "grad_norm": 0.8214516639709473, |
| "learning_rate": 4.7761194029850745e-06, |
| "loss": 0.6202, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.14823133071308253, |
| "grad_norm": 0.7563803195953369, |
| "learning_rate": 4.925373134328359e-06, |
| "loss": 0.6047, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.1527231892195396, |
| "grad_norm": 0.5525519251823425, |
| "learning_rate": 5.074626865671642e-06, |
| "loss": 0.6116, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.15721504772599662, |
| "grad_norm": 0.5453880429267883, |
| "learning_rate": 5.2238805970149255e-06, |
| "loss": 0.6055, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.16170690623245368, |
| "grad_norm": 0.6540727615356445, |
| "learning_rate": 5.37313432835821e-06, |
| "loss": 0.5947, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.16619876473891074, |
| "grad_norm": 0.650087833404541, |
| "learning_rate": 5.522388059701493e-06, |
| "loss": 0.5908, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.17069062324536777, |
| "grad_norm": 0.5679633617401123, |
| "learning_rate": 5.671641791044776e-06, |
| "loss": 0.598, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.17518248175182483, |
| "grad_norm": 0.44914695620536804, |
| "learning_rate": 5.820895522388061e-06, |
| "loss": 0.5927, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.17967434025828186, |
| "grad_norm": 0.47071295976638794, |
| "learning_rate": 5.970149253731343e-06, |
| "loss": 0.596, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.18416619876473891, |
| "grad_norm": 0.5729812979698181, |
| "learning_rate": 6.119402985074627e-06, |
| "loss": 0.5919, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.18865805727119594, |
| "grad_norm": 0.5159889459609985, |
| "learning_rate": 6.2686567164179116e-06, |
| "loss": 0.5767, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.193149915777653, |
| "grad_norm": 0.39863625168800354, |
| "learning_rate": 6.417910447761194e-06, |
| "loss": 0.5741, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.19764177428411006, |
| "grad_norm": 0.46757975220680237, |
| "learning_rate": 6.567164179104478e-06, |
| "loss": 0.5813, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.2021336327905671, |
| "grad_norm": 0.43369781970977783, |
| "learning_rate": 6.7164179104477625e-06, |
| "loss": 0.5664, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.20662549129702415, |
| "grad_norm": 0.35700419545173645, |
| "learning_rate": 6.865671641791045e-06, |
| "loss": 0.5718, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.21111734980348118, |
| "grad_norm": 0.4011654853820801, |
| "learning_rate": 7.014925373134329e-06, |
| "loss": 0.5637, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.21560920830993824, |
| "grad_norm": 0.43147680163383484, |
| "learning_rate": 7.164179104477612e-06, |
| "loss": 0.5688, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.2201010668163953, |
| "grad_norm": 0.41290047764778137, |
| "learning_rate": 7.313432835820896e-06, |
| "loss": 0.5554, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.22459292532285233, |
| "grad_norm": 0.40266507863998413, |
| "learning_rate": 7.46268656716418e-06, |
| "loss": 0.5759, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.22908478382930939, |
| "grad_norm": 0.30240580439567566, |
| "learning_rate": 7.611940298507463e-06, |
| "loss": 0.5586, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.23357664233576642, |
| "grad_norm": 0.31717273592948914, |
| "learning_rate": 7.761194029850747e-06, |
| "loss": 0.5639, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.23806850084222347, |
| "grad_norm": 0.35987257957458496, |
| "learning_rate": 7.91044776119403e-06, |
| "loss": 0.5561, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.2425603593486805, |
| "grad_norm": 0.3291323781013489, |
| "learning_rate": 8.059701492537314e-06, |
| "loss": 0.5595, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.24705221785513756, |
| "grad_norm": 0.28238141536712646, |
| "learning_rate": 8.208955223880599e-06, |
| "loss": 0.5374, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.2515440763615946, |
| "grad_norm": 0.30300024151802063, |
| "learning_rate": 8.35820895522388e-06, |
| "loss": 0.5514, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.25603593486805165, |
| "grad_norm": 0.27751174569129944, |
| "learning_rate": 8.507462686567165e-06, |
| "loss": 0.5465, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.2605277933745087, |
| "grad_norm": 0.24813084304332733, |
| "learning_rate": 8.656716417910447e-06, |
| "loss": 0.5388, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.26501965188096577, |
| "grad_norm": 0.2552523612976074, |
| "learning_rate": 8.805970149253732e-06, |
| "loss": 0.553, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.26951151038742277, |
| "grad_norm": 0.2615125775337219, |
| "learning_rate": 8.955223880597016e-06, |
| "loss": 0.5414, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.27400336889387983, |
| "grad_norm": 0.2671511173248291, |
| "learning_rate": 9.104477611940299e-06, |
| "loss": 0.5356, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.2784952274003369, |
| "grad_norm": 0.2728058397769928, |
| "learning_rate": 9.253731343283582e-06, |
| "loss": 0.5413, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.28298708590679394, |
| "grad_norm": 0.2823947072029114, |
| "learning_rate": 9.402985074626867e-06, |
| "loss": 0.5505, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.287478944413251, |
| "grad_norm": 0.28685927391052246, |
| "learning_rate": 9.552238805970149e-06, |
| "loss": 0.5346, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.291970802919708, |
| "grad_norm": 0.24975822865962982, |
| "learning_rate": 9.701492537313434e-06, |
| "loss": 0.5366, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.29646266142616506, |
| "grad_norm": 0.2949586808681488, |
| "learning_rate": 9.850746268656717e-06, |
| "loss": 0.5247, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.3009545199326221, |
| "grad_norm": 0.2750069797039032, |
| "learning_rate": 1e-05, |
| "loss": 0.538, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.3054463784390792, |
| "grad_norm": 0.34335073828697205, |
| "learning_rate": 9.999931232202689e-06, |
| "loss": 0.5252, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.30993823694553624, |
| "grad_norm": 0.24552962183952332, |
| "learning_rate": 9.999724930702358e-06, |
| "loss": 0.5312, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.31443009545199324, |
| "grad_norm": 0.3122241795063019, |
| "learning_rate": 9.999381101173765e-06, |
| "loss": 0.5281, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.3189219539584503, |
| "grad_norm": 0.23177407681941986, |
| "learning_rate": 9.99889975307467e-06, |
| "loss": 0.5266, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.32341381246490736, |
| "grad_norm": 0.2342972457408905, |
| "learning_rate": 9.998280899645575e-06, |
| "loss": 0.5419, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.3279056709713644, |
| "grad_norm": 0.26972994208335876, |
| "learning_rate": 9.997524557909353e-06, |
| "loss": 0.528, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.3323975294778215, |
| "grad_norm": 0.2486950308084488, |
| "learning_rate": 9.996630748670788e-06, |
| "loss": 0.5285, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.3368893879842785, |
| "grad_norm": 0.2598794400691986, |
| "learning_rate": 9.995599496515996e-06, |
| "loss": 0.5339, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.34138124649073553, |
| "grad_norm": 0.2780413329601288, |
| "learning_rate": 9.99443082981175e-06, |
| "loss": 0.5209, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.3458731049971926, |
| "grad_norm": 0.23097486793994904, |
| "learning_rate": 9.993124780704707e-06, |
| "loss": 0.534, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.35036496350364965, |
| "grad_norm": 0.23875422775745392, |
| "learning_rate": 9.991681385120515e-06, |
| "loss": 0.5152, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.35485682201010665, |
| "grad_norm": 0.28855064511299133, |
| "learning_rate": 9.99010068276283e-06, |
| "loss": 0.5279, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.3593486805165637, |
| "grad_norm": 0.24532583355903625, |
| "learning_rate": 9.988382717112213e-06, |
| "loss": 0.53, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.36384053902302077, |
| "grad_norm": 0.2871251404285431, |
| "learning_rate": 9.986527535424956e-06, |
| "loss": 0.529, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.36833239752947783, |
| "grad_norm": 0.23627132177352905, |
| "learning_rate": 9.98453518873176e-06, |
| "loss": 0.5331, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.3728242560359349, |
| "grad_norm": 0.2671157419681549, |
| "learning_rate": 9.982405731836343e-06, |
| "loss": 0.5241, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.3773161145423919, |
| "grad_norm": 0.2739470899105072, |
| "learning_rate": 9.980139223313926e-06, |
| "loss": 0.5232, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.38180797304884895, |
| "grad_norm": 0.2618931531906128, |
| "learning_rate": 9.977735725509632e-06, |
| "loss": 0.5261, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.386299831555306, |
| "grad_norm": 0.29810863733291626, |
| "learning_rate": 9.97519530453676e-06, |
| "loss": 0.5089, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.39079169006176306, |
| "grad_norm": 0.26110532879829407, |
| "learning_rate": 9.97251803027497e-06, |
| "loss": 0.5082, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.3952835485682201, |
| "grad_norm": 0.23007212579250336, |
| "learning_rate": 9.969703976368368e-06, |
| "loss": 0.5128, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.3997754070746771, |
| "grad_norm": 0.2656469941139221, |
| "learning_rate": 9.966753220223466e-06, |
| "loss": 0.5213, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.4042672655811342, |
| "grad_norm": 0.2582208812236786, |
| "learning_rate": 9.963665843007066e-06, |
| "loss": 0.5192, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.40875912408759124, |
| "grad_norm": 0.24596616625785828, |
| "learning_rate": 9.960441929644017e-06, |
| "loss": 0.52, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.4132509825940483, |
| "grad_norm": 0.22645823657512665, |
| "learning_rate": 9.95708156881489e-06, |
| "loss": 0.5114, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.41774284110050536, |
| "grad_norm": 0.2343415766954422, |
| "learning_rate": 9.95358485295353e-06, |
| "loss": 0.5124, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.42223469960696236, |
| "grad_norm": 0.22430844604969025, |
| "learning_rate": 9.949951878244514e-06, |
| "loss": 0.5202, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.4267265581134194, |
| "grad_norm": 0.2198512703180313, |
| "learning_rate": 9.946182744620512e-06, |
| "loss": 0.5198, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.4312184166198765, |
| "grad_norm": 0.2237306386232376, |
| "learning_rate": 9.94227755575953e-06, |
| "loss": 0.4995, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.43571027512633353, |
| "grad_norm": 0.2366565316915512, |
| "learning_rate": 9.93823641908206e-06, |
| "loss": 0.5182, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.4402021336327906, |
| "grad_norm": 0.24728058278560638, |
| "learning_rate": 9.934059445748134e-06, |
| "loss": 0.5124, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.4446939921392476, |
| "grad_norm": 0.22210274636745453, |
| "learning_rate": 9.92974675065425e-06, |
| "loss": 0.4915, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.44918585064570465, |
| "grad_norm": 0.22185443341732025, |
| "learning_rate": 9.925298452430225e-06, |
| "loss": 0.5033, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.4536777091521617, |
| "grad_norm": 0.278481662273407, |
| "learning_rate": 9.920714673435931e-06, |
| "loss": 0.5156, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.45816956765861877, |
| "grad_norm": 0.2632293999195099, |
| "learning_rate": 9.915995539757918e-06, |
| "loss": 0.5114, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.4626614261650758, |
| "grad_norm": 0.29412922263145447, |
| "learning_rate": 9.91114118120596e-06, |
| "loss": 0.5117, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.46715328467153283, |
| "grad_norm": 0.2760590612888336, |
| "learning_rate": 9.906151731309472e-06, |
| "loss": 0.5013, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.4716451431779899, |
| "grad_norm": 0.2319260984659195, |
| "learning_rate": 9.901027327313847e-06, |
| "loss": 0.4969, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.47613700168444695, |
| "grad_norm": 0.22962868213653564, |
| "learning_rate": 9.895768110176677e-06, |
| "loss": 0.5045, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.480628860190904, |
| "grad_norm": 0.2626447081565857, |
| "learning_rate": 9.890374224563872e-06, |
| "loss": 0.5033, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.485120718697361, |
| "grad_norm": 0.2323921024799347, |
| "learning_rate": 9.884845818845685e-06, |
| "loss": 0.5136, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.48961257720381807, |
| "grad_norm": 0.32458624243736267, |
| "learning_rate": 9.879183045092628e-06, |
| "loss": 0.5142, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.4941044357102751, |
| "grad_norm": 0.23757074773311615, |
| "learning_rate": 9.873386059071294e-06, |
| "loss": 0.5086, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.4985962942167322, |
| "grad_norm": 0.27966201305389404, |
| "learning_rate": 9.86745502024007e-06, |
| "loss": 0.5095, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.5030881527231892, |
| "grad_norm": 0.2681194543838501, |
| "learning_rate": 9.861390091744738e-06, |
| "loss": 0.4988, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.5075800112296462, |
| "grad_norm": 0.24083542823791504, |
| "learning_rate": 9.855191440414014e-06, |
| "loss": 0.5152, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.5120718697361033, |
| "grad_norm": 0.2831341326236725, |
| "learning_rate": 9.848859236754936e-06, |
| "loss": 0.5098, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.5165637282425604, |
| "grad_norm": 0.27346932888031006, |
| "learning_rate": 9.84239365494818e-06, |
| "loss": 0.497, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.5210555867490174, |
| "grad_norm": 0.27766865491867065, |
| "learning_rate": 9.835794872843281e-06, |
| "loss": 0.5183, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.5255474452554745, |
| "grad_norm": 0.26060062646865845, |
| "learning_rate": 9.829063071953715e-06, |
| "loss": 0.5031, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.5300393037619315, |
| "grad_norm": 0.2610337734222412, |
| "learning_rate": 9.822198437451933e-06, |
| "loss": 0.5094, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.5345311622683886, |
| "grad_norm": 0.2221408635377884, |
| "learning_rate": 9.815201158164253e-06, |
| "loss": 0.5015, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.5390230207748455, |
| "grad_norm": 0.2650430202484131, |
| "learning_rate": 9.808071426565671e-06, |
| "loss": 0.5021, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.5435148792813026, |
| "grad_norm": 0.2397095412015915, |
| "learning_rate": 9.800809438774557e-06, |
| "loss": 0.5096, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.5480067377877597, |
| "grad_norm": 0.3001818358898163, |
| "learning_rate": 9.793415394547274e-06, |
| "loss": 0.5019, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.5524985962942167, |
| "grad_norm": 0.22701479494571686, |
| "learning_rate": 9.785889497272678e-06, |
| "loss": 0.4856, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.5569904548006738, |
| "grad_norm": 0.32053694128990173, |
| "learning_rate": 9.778231953966519e-06, |
| "loss": 0.5132, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.5614823133071308, |
| "grad_norm": 0.2665109932422638, |
| "learning_rate": 9.770442975265753e-06, |
| "loss": 0.5144, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.5659741718135879, |
| "grad_norm": 0.2617216408252716, |
| "learning_rate": 9.762522775422741e-06, |
| "loss": 0.4918, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.570466030320045, |
| "grad_norm": 0.27840656042099, |
| "learning_rate": 9.754471572299363e-06, |
| "loss": 0.494, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.574957888826502, |
| "grad_norm": 0.2666041851043701, |
| "learning_rate": 9.746289587361021e-06, |
| "loss": 0.4986, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.5794497473329591, |
| "grad_norm": 0.27760958671569824, |
| "learning_rate": 9.737977045670549e-06, |
| "loss": 0.5042, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.583941605839416, |
| "grad_norm": 0.2891870141029358, |
| "learning_rate": 9.729534175882016e-06, |
| "loss": 0.4888, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.5884334643458731, |
| "grad_norm": 0.24741241335868835, |
| "learning_rate": 9.72096121023445e-06, |
| "loss": 0.5037, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.5929253228523301, |
| "grad_norm": 0.3258378803730011, |
| "learning_rate": 9.712258384545432e-06, |
| "loss": 0.4874, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.5974171813587872, |
| "grad_norm": 0.26103854179382324, |
| "learning_rate": 9.703425938204627e-06, |
| "loss": 0.4928, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.6019090398652442, |
| "grad_norm": 0.2776138186454773, |
| "learning_rate": 9.694464114167185e-06, |
| "loss": 0.4993, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.6064008983717013, |
| "grad_norm": 0.2927091717720032, |
| "learning_rate": 9.685373158947067e-06, |
| "loss": 0.4993, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.6108927568781584, |
| "grad_norm": 0.31002023816108704, |
| "learning_rate": 9.676153322610259e-06, |
| "loss": 0.5032, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.6153846153846154, |
| "grad_norm": 0.31717565655708313, |
| "learning_rate": 9.666804858767894e-06, |
| "loss": 0.4964, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.6198764738910725, |
| "grad_norm": 0.2911582887172699, |
| "learning_rate": 9.65732802456928e-06, |
| "loss": 0.4929, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.6243683323975294, |
| "grad_norm": 0.315828800201416, |
| "learning_rate": 9.647723080694822e-06, |
| "loss": 0.4988, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.6288601909039865, |
| "grad_norm": 0.2951522171497345, |
| "learning_rate": 9.637990291348853e-06, |
| "loss": 0.4908, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.6333520494104435, |
| "grad_norm": 0.26586809754371643, |
| "learning_rate": 9.628129924252368e-06, |
| "loss": 0.5017, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.6378439079169006, |
| "grad_norm": 0.33742159605026245, |
| "learning_rate": 9.618142250635658e-06, |
| "loss": 0.4917, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.6423357664233577, |
| "grad_norm": 0.24248960614204407, |
| "learning_rate": 9.608027545230847e-06, |
| "loss": 0.4844, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.6468276249298147, |
| "grad_norm": 0.26963257789611816, |
| "learning_rate": 9.597786086264339e-06, |
| "loss": 0.4869, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.6513194834362718, |
| "grad_norm": 0.2914552390575409, |
| "learning_rate": 9.587418155449167e-06, |
| "loss": 0.503, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.6558113419427288, |
| "grad_norm": 0.2754093408584595, |
| "learning_rate": 9.576924037977233e-06, |
| "loss": 0.496, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.6603032004491859, |
| "grad_norm": 0.278185099363327, |
| "learning_rate": 9.566304022511477e-06, |
| "loss": 0.4944, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.664795058955643, |
| "grad_norm": 0.24680428206920624, |
| "learning_rate": 9.555558401177927e-06, |
| "loss": 0.4931, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.6692869174620999, |
| "grad_norm": 0.29219385981559753, |
| "learning_rate": 9.544687469557667e-06, |
| "loss": 0.4965, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.673778775968557, |
| "grad_norm": 0.27283161878585815, |
| "learning_rate": 9.533691526678705e-06, |
| "loss": 0.4893, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.678270634475014, |
| "grad_norm": 0.2558472156524658, |
| "learning_rate": 9.52257087500775e-06, |
| "loss": 0.5045, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.6827624929814711, |
| "grad_norm": 0.2973162829875946, |
| "learning_rate": 9.51132582044189e-06, |
| "loss": 0.512, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.6872543514879281, |
| "grad_norm": 0.3723624050617218, |
| "learning_rate": 9.49995667230018e-06, |
| "loss": 0.4986, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.6917462099943852, |
| "grad_norm": 0.24720445275306702, |
| "learning_rate": 9.488463743315126e-06, |
| "loss": 0.4925, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.6962380685008422, |
| "grad_norm": 0.37346410751342773, |
| "learning_rate": 9.476847349624097e-06, |
| "loss": 0.4974, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.7007299270072993, |
| "grad_norm": 0.31400275230407715, |
| "learning_rate": 9.46510781076061e-06, |
| "loss": 0.4942, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.7052217855137564, |
| "grad_norm": 0.29658353328704834, |
| "learning_rate": 9.453245449645562e-06, |
| "loss": 0.4865, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.7097136440202133, |
| "grad_norm": 0.2616643011569977, |
| "learning_rate": 9.44126059257833e-06, |
| "loss": 0.4932, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.7142055025266704, |
| "grad_norm": 0.2881576716899872, |
| "learning_rate": 9.4291535692278e-06, |
| "loss": 0.5062, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.7186973610331274, |
| "grad_norm": 0.24668999016284943, |
| "learning_rate": 9.416924712623305e-06, |
| "loss": 0.4994, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.7231892195395845, |
| "grad_norm": 0.2936095893383026, |
| "learning_rate": 9.40457435914546e-06, |
| "loss": 0.4885, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.7276810780460415, |
| "grad_norm": 0.25287431478500366, |
| "learning_rate": 9.392102848516901e-06, |
| "loss": 0.4906, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.7321729365524986, |
| "grad_norm": 0.2542520761489868, |
| "learning_rate": 9.37951052379296e-06, |
| "loss": 0.4747, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.7366647950589557, |
| "grad_norm": 0.25072625279426575, |
| "learning_rate": 9.36679773135221e-06, |
| "loss": 0.4799, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.7411566535654127, |
| "grad_norm": 0.24300134181976318, |
| "learning_rate": 9.353964820886938e-06, |
| "loss": 0.4963, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.7456485120718698, |
| "grad_norm": 0.24435429275035858, |
| "learning_rate": 9.341012145393546e-06, |
| "loss": 0.4961, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.7501403705783268, |
| "grad_norm": 0.27164775133132935, |
| "learning_rate": 9.327940061162816e-06, |
| "loss": 0.5013, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.7546322290847838, |
| "grad_norm": 0.24927161633968353, |
| "learning_rate": 9.314748927770126e-06, |
| "loss": 0.4843, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.7591240875912408, |
| "grad_norm": 0.2827945649623871, |
| "learning_rate": 9.301439108065546e-06, |
| "loss": 0.4902, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.7636159460976979, |
| "grad_norm": 0.22944357991218567, |
| "learning_rate": 9.288010968163874e-06, |
| "loss": 0.4886, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.768107804604155, |
| "grad_norm": 0.2653137445449829, |
| "learning_rate": 9.274464877434548e-06, |
| "loss": 0.4974, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.772599663110612, |
| "grad_norm": 0.26817506551742554, |
| "learning_rate": 9.260801208491499e-06, |
| "loss": 0.4878, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.7770915216170691, |
| "grad_norm": 0.21729539334774017, |
| "learning_rate": 9.247020337182892e-06, |
| "loss": 0.4778, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.7815833801235261, |
| "grad_norm": 0.2882274091243744, |
| "learning_rate": 9.233122642580796e-06, |
| "loss": 0.4924, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.7860752386299832, |
| "grad_norm": 0.2678658962249756, |
| "learning_rate": 9.219108506970747e-06, |
| "loss": 0.496, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.7905670971364402, |
| "grad_norm": 0.3008089065551758, |
| "learning_rate": 9.204978315841238e-06, |
| "loss": 0.4947, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.7950589556428973, |
| "grad_norm": 0.25311562418937683, |
| "learning_rate": 9.19073245787312e-06, |
| "loss": 0.4751, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.7995508141493542, |
| "grad_norm": 0.31886932253837585, |
| "learning_rate": 9.1763713249289e-06, |
| "loss": 0.4831, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.8040426726558113, |
| "grad_norm": 0.2668146789073944, |
| "learning_rate": 9.16189531204197e-06, |
| "loss": 0.4866, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.8085345311622684, |
| "grad_norm": 0.25939807295799255, |
| "learning_rate": 9.147304817405741e-06, |
| "loss": 0.4928, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.8130263896687254, |
| "grad_norm": 0.27832671999931335, |
| "learning_rate": 9.132600242362682e-06, |
| "loss": 0.5013, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.8175182481751825, |
| "grad_norm": 0.24878981709480286, |
| "learning_rate": 9.117781991393285e-06, |
| "loss": 0.4839, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.8220101066816395, |
| "grad_norm": 0.2747175097465515, |
| "learning_rate": 9.102850472104945e-06, |
| "loss": 0.4888, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.8265019651880966, |
| "grad_norm": 0.2612956762313843, |
| "learning_rate": 9.08780609522074e-06, |
| "loss": 0.481, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.8309938236945537, |
| "grad_norm": 0.2580430209636688, |
| "learning_rate": 9.07264927456813e-06, |
| "loss": 0.4806, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.8354856822010107, |
| "grad_norm": 0.31817421317100525, |
| "learning_rate": 9.057380427067584e-06, |
| "loss": 0.4864, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.8399775407074677, |
| "grad_norm": 0.28587251901626587, |
| "learning_rate": 9.04199997272111e-06, |
| "loss": 0.4882, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.8444693992139247, |
| "grad_norm": 0.2851974368095398, |
| "learning_rate": 9.02650833460069e-06, |
| "loss": 0.4786, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.8489612577203818, |
| "grad_norm": 0.2742379605770111, |
| "learning_rate": 9.01090593883666e-06, |
| "loss": 0.4793, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.8534531162268388, |
| "grad_norm": 0.29192987084388733, |
| "learning_rate": 8.995193214605972e-06, |
| "loss": 0.495, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.8579449747332959, |
| "grad_norm": 0.3218548893928528, |
| "learning_rate": 8.979370594120404e-06, |
| "loss": 0.4962, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.862436833239753, |
| "grad_norm": 0.21388249099254608, |
| "learning_rate": 8.963438512614657e-06, |
| "loss": 0.4923, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.86692869174621, |
| "grad_norm": 0.29332301020622253, |
| "learning_rate": 8.94739740833439e-06, |
| "loss": 0.4883, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.8714205502526671, |
| "grad_norm": 0.2745992839336395, |
| "learning_rate": 8.93124772252417e-06, |
| "loss": 0.4992, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.8759124087591241, |
| "grad_norm": 0.22587467730045319, |
| "learning_rate": 8.914989899415323e-06, |
| "loss": 0.4741, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.8804042672655812, |
| "grad_norm": 0.2980908453464508, |
| "learning_rate": 8.898624386213724e-06, |
| "loss": 0.4777, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.8848961257720381, |
| "grad_norm": 0.3256934881210327, |
| "learning_rate": 8.88215163308749e-06, |
| "loss": 0.4859, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.8893879842784952, |
| "grad_norm": 0.27209722995758057, |
| "learning_rate": 8.8655720931546e-06, |
| "loss": 0.4948, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.8938798427849522, |
| "grad_norm": 0.26153564453125, |
| "learning_rate": 8.84888622247043e-06, |
| "loss": 0.4844, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.8983717012914093, |
| "grad_norm": 0.3023647367954254, |
| "learning_rate": 8.832094480015211e-06, |
| "loss": 0.4711, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.9028635597978664, |
| "grad_norm": 0.2709749937057495, |
| "learning_rate": 8.815197327681399e-06, |
| "loss": 0.4839, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.9073554183043234, |
| "grad_norm": 0.2602469325065613, |
| "learning_rate": 8.798195230260973e-06, |
| "loss": 0.5001, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.9118472768107805, |
| "grad_norm": 0.2420167326927185, |
| "learning_rate": 8.781088655432648e-06, |
| "loss": 0.4774, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.9163391353172375, |
| "grad_norm": 0.2579791843891144, |
| "learning_rate": 8.763878073749013e-06, |
| "loss": 0.4784, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.9208309938236946, |
| "grad_norm": 0.24879693984985352, |
| "learning_rate": 8.746563958623584e-06, |
| "loss": 0.4879, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.9253228523301515, |
| "grad_norm": 0.24991406500339508, |
| "learning_rate": 8.729146786317787e-06, |
| "loss": 0.4931, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.9298147108366086, |
| "grad_norm": 0.285593181848526, |
| "learning_rate": 8.711627035927848e-06, |
| "loss": 0.4815, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.9343065693430657, |
| "grad_norm": 0.2333066463470459, |
| "learning_rate": 8.694005189371627e-06, |
| "loss": 0.4794, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.9387984278495227, |
| "grad_norm": 0.271056592464447, |
| "learning_rate": 8.676281731375355e-06, |
| "loss": 0.4893, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.9432902863559798, |
| "grad_norm": 0.24806849658489227, |
| "learning_rate": 8.658457149460296e-06, |
| "loss": 0.4929, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.9477821448624368, |
| "grad_norm": 0.25548022985458374, |
| "learning_rate": 8.640531933929344e-06, |
| "loss": 0.4828, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.9522740033688939, |
| "grad_norm": 0.2774530053138733, |
| "learning_rate": 8.622506577853538e-06, |
| "loss": 0.4764, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.956765861875351, |
| "grad_norm": 0.2682695984840393, |
| "learning_rate": 8.604381577058486e-06, |
| "loss": 0.476, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.961257720381808, |
| "grad_norm": 0.24734169244766235, |
| "learning_rate": 8.586157430110747e-06, |
| "loss": 0.4858, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.9657495788882651, |
| "grad_norm": 0.2741633951663971, |
| "learning_rate": 8.56783463830409e-06, |
| "loss": 0.4647, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.970241437394722, |
| "grad_norm": 0.2486443966627121, |
| "learning_rate": 8.549413705645737e-06, |
| "loss": 0.481, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.9747332959011791, |
| "grad_norm": 0.25736650824546814, |
| "learning_rate": 8.530895138842467e-06, |
| "loss": 0.4764, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.9792251544076361, |
| "grad_norm": 0.22502939403057098, |
| "learning_rate": 8.512279447286704e-06, |
| "loss": 0.484, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.9837170129140932, |
| "grad_norm": 0.24173305928707123, |
| "learning_rate": 8.493567143042485e-06, |
| "loss": 0.4827, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.9882088714205502, |
| "grad_norm": 0.20804060995578766, |
| "learning_rate": 8.47475874083139e-06, |
| "loss": 0.4963, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.9927007299270073, |
| "grad_norm": 0.24134287238121033, |
| "learning_rate": 8.455854758018377e-06, |
| "loss": 0.4844, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.9971925884334644, |
| "grad_norm": 0.23423220217227936, |
| "learning_rate": 8.436855714597548e-06, |
| "loss": 0.4819, |
| "step": 222 |
| }, |
| { |
| "epoch": 1.0030881527231892, |
| "grad_norm": 0.4805089831352234, |
| "learning_rate": 8.417762133177849e-06, |
| "loss": 0.7888, |
| "step": 223 |
| }, |
| { |
| "epoch": 1.0075800112296462, |
| "grad_norm": 0.43639126420021057, |
| "learning_rate": 8.398574538968697e-06, |
| "loss": 0.4816, |
| "step": 224 |
| }, |
| { |
| "epoch": 1.0120718697361033, |
| "grad_norm": 0.36428216099739075, |
| "learning_rate": 8.379293459765527e-06, |
| "loss": 0.4509, |
| "step": 225 |
| }, |
| { |
| "epoch": 1.0165637282425604, |
| "grad_norm": 0.33154451847076416, |
| "learning_rate": 8.359919425935276e-06, |
| "loss": 0.4652, |
| "step": 226 |
| }, |
| { |
| "epoch": 1.0210555867490174, |
| "grad_norm": 0.28618019819259644, |
| "learning_rate": 8.340452970401798e-06, |
| "loss": 0.462, |
| "step": 227 |
| }, |
| { |
| "epoch": 1.0255474452554745, |
| "grad_norm": 0.35930895805358887, |
| "learning_rate": 8.3208946286312e-06, |
| "loss": 0.485, |
| "step": 228 |
| }, |
| { |
| "epoch": 1.0300393037619315, |
| "grad_norm": 0.29257896542549133, |
| "learning_rate": 8.301244938617117e-06, |
| "loss": 0.4649, |
| "step": 229 |
| }, |
| { |
| "epoch": 1.0345311622683886, |
| "grad_norm": 0.28928303718566895, |
| "learning_rate": 8.281504440865905e-06, |
| "loss": 0.4828, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.0390230207748457, |
| "grad_norm": 0.3513385057449341, |
| "learning_rate": 8.261673678381786e-06, |
| "loss": 0.4787, |
| "step": 231 |
| }, |
| { |
| "epoch": 1.0435148792813027, |
| "grad_norm": 0.334462970495224, |
| "learning_rate": 8.241753196651903e-06, |
| "loss": 0.4859, |
| "step": 232 |
| }, |
| { |
| "epoch": 1.0480067377877598, |
| "grad_norm": 0.2647038996219635, |
| "learning_rate": 8.221743543631314e-06, |
| "loss": 0.4701, |
| "step": 233 |
| }, |
| { |
| "epoch": 1.0524985962942168, |
| "grad_norm": 0.36219117045402527, |
| "learning_rate": 8.201645269727924e-06, |
| "loss": 0.4675, |
| "step": 234 |
| }, |
| { |
| "epoch": 1.0569904548006739, |
| "grad_norm": 0.3206924796104431, |
| "learning_rate": 8.181458927787347e-06, |
| "loss": 0.4782, |
| "step": 235 |
| }, |
| { |
| "epoch": 1.0614823133071307, |
| "grad_norm": 0.29768282175064087, |
| "learning_rate": 8.161185073077686e-06, |
| "loss": 0.4695, |
| "step": 236 |
| }, |
| { |
| "epoch": 1.0659741718135878, |
| "grad_norm": 0.32597997784614563, |
| "learning_rate": 8.140824263274278e-06, |
| "loss": 0.4546, |
| "step": 237 |
| }, |
| { |
| "epoch": 1.0704660303200448, |
| "grad_norm": 0.26331695914268494, |
| "learning_rate": 8.120377058444336e-06, |
| "loss": 0.4749, |
| "step": 238 |
| }, |
| { |
| "epoch": 1.074957888826502, |
| "grad_norm": 0.2784285843372345, |
| "learning_rate": 8.099844021031559e-06, |
| "loss": 0.4664, |
| "step": 239 |
| }, |
| { |
| "epoch": 1.079449747332959, |
| "grad_norm": 0.30415502190589905, |
| "learning_rate": 8.079225715840646e-06, |
| "loss": 0.4611, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.083941605839416, |
| "grad_norm": 0.2318006455898285, |
| "learning_rate": 8.058522710021773e-06, |
| "loss": 0.4647, |
| "step": 241 |
| }, |
| { |
| "epoch": 1.088433464345873, |
| "grad_norm": 0.3010203242301941, |
| "learning_rate": 8.037735573054979e-06, |
| "loss": 0.4778, |
| "step": 242 |
| }, |
| { |
| "epoch": 1.0929253228523301, |
| "grad_norm": 0.29548323154449463, |
| "learning_rate": 8.016864876734514e-06, |
| "loss": 0.4679, |
| "step": 243 |
| }, |
| { |
| "epoch": 1.0974171813587872, |
| "grad_norm": 0.2665966749191284, |
| "learning_rate": 7.995911195153105e-06, |
| "loss": 0.4559, |
| "step": 244 |
| }, |
| { |
| "epoch": 1.1019090398652442, |
| "grad_norm": 0.2814834713935852, |
| "learning_rate": 7.974875104686164e-06, |
| "loss": 0.4668, |
| "step": 245 |
| }, |
| { |
| "epoch": 1.1064008983717013, |
| "grad_norm": 0.3019580841064453, |
| "learning_rate": 7.95375718397593e-06, |
| "loss": 0.4915, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.1108927568781584, |
| "grad_norm": 0.2376527190208435, |
| "learning_rate": 7.932558013915561e-06, |
| "loss": 0.4494, |
| "step": 247 |
| }, |
| { |
| "epoch": 1.1153846153846154, |
| "grad_norm": 0.24670954048633575, |
| "learning_rate": 7.911278177633151e-06, |
| "loss": 0.468, |
| "step": 248 |
| }, |
| { |
| "epoch": 1.1198764738910725, |
| "grad_norm": 0.25502070784568787, |
| "learning_rate": 7.889918260475685e-06, |
| "loss": 0.4723, |
| "step": 249 |
| }, |
| { |
| "epoch": 1.1243683323975295, |
| "grad_norm": 0.23503544926643372, |
| "learning_rate": 7.868478849992944e-06, |
| "loss": 0.4913, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.1288601909039866, |
| "grad_norm": 0.2687453031539917, |
| "learning_rate": 7.846960535921344e-06, |
| "loss": 0.468, |
| "step": 251 |
| }, |
| { |
| "epoch": 1.1333520494104437, |
| "grad_norm": 0.2521950304508209, |
| "learning_rate": 7.825363910167709e-06, |
| "loss": 0.4421, |
| "step": 252 |
| }, |
| { |
| "epoch": 1.1378439079169007, |
| "grad_norm": 0.240973562002182, |
| "learning_rate": 7.803689566792989e-06, |
| "loss": 0.4902, |
| "step": 253 |
| }, |
| { |
| "epoch": 1.1423357664233578, |
| "grad_norm": 0.254777729511261, |
| "learning_rate": 7.781938101995928e-06, |
| "loss": 0.4518, |
| "step": 254 |
| }, |
| { |
| "epoch": 1.1468276249298146, |
| "grad_norm": 0.2425854653120041, |
| "learning_rate": 7.76011011409665e-06, |
| "loss": 0.4836, |
| "step": 255 |
| }, |
| { |
| "epoch": 1.1513194834362717, |
| "grad_norm": 0.256160169839859, |
| "learning_rate": 7.738206203520223e-06, |
| "loss": 0.4478, |
| "step": 256 |
| }, |
| { |
| "epoch": 1.1558113419427287, |
| "grad_norm": 0.2657393515110016, |
| "learning_rate": 7.716226972780111e-06, |
| "loss": 0.4838, |
| "step": 257 |
| }, |
| { |
| "epoch": 1.1603032004491858, |
| "grad_norm": 0.2680897116661072, |
| "learning_rate": 7.694173026461634e-06, |
| "loss": 0.465, |
| "step": 258 |
| }, |
| { |
| "epoch": 1.1647950589556428, |
| "grad_norm": 0.37795037031173706, |
| "learning_rate": 7.672044971205315e-06, |
| "loss": 0.481, |
| "step": 259 |
| }, |
| { |
| "epoch": 1.1692869174621, |
| "grad_norm": 0.24703949689865112, |
| "learning_rate": 7.649843415690198e-06, |
| "loss": 0.4517, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.173778775968557, |
| "grad_norm": 0.3117562532424927, |
| "learning_rate": 7.627568970617114e-06, |
| "loss": 0.4445, |
| "step": 261 |
| }, |
| { |
| "epoch": 1.178270634475014, |
| "grad_norm": 0.2995341420173645, |
| "learning_rate": 7.6052222486918725e-06, |
| "loss": 0.4928, |
| "step": 262 |
| }, |
| { |
| "epoch": 1.182762492981471, |
| "grad_norm": 0.2563771903514862, |
| "learning_rate": 7.582803864608411e-06, |
| "loss": 0.462, |
| "step": 263 |
| }, |
| { |
| "epoch": 1.1872543514879281, |
| "grad_norm": 0.30361172556877136, |
| "learning_rate": 7.560314435031886e-06, |
| "loss": 0.4723, |
| "step": 264 |
| }, |
| { |
| "epoch": 1.1917462099943852, |
| "grad_norm": 0.24592924118041992, |
| "learning_rate": 7.537754578581711e-06, |
| "loss": 0.4569, |
| "step": 265 |
| }, |
| { |
| "epoch": 1.1962380685008422, |
| "grad_norm": 0.28482651710510254, |
| "learning_rate": 7.51512491581454e-06, |
| "loss": 0.474, |
| "step": 266 |
| }, |
| { |
| "epoch": 1.2007299270072993, |
| "grad_norm": 0.28539368510246277, |
| "learning_rate": 7.4924260692072e-06, |
| "loss": 0.4665, |
| "step": 267 |
| }, |
| { |
| "epoch": 1.2052217855137564, |
| "grad_norm": 0.2118852585554123, |
| "learning_rate": 7.4696586631395626e-06, |
| "loss": 0.4665, |
| "step": 268 |
| }, |
| { |
| "epoch": 1.2097136440202134, |
| "grad_norm": 0.27248629927635193, |
| "learning_rate": 7.446823323877375e-06, |
| "loss": 0.4749, |
| "step": 269 |
| }, |
| { |
| "epoch": 1.2142055025266705, |
| "grad_norm": 0.24900004267692566, |
| "learning_rate": 7.423920679555029e-06, |
| "loss": 0.4492, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.2186973610331275, |
| "grad_norm": 0.23359505832195282, |
| "learning_rate": 7.400951360158285e-06, |
| "loss": 0.4866, |
| "step": 271 |
| }, |
| { |
| "epoch": 1.2231892195395846, |
| "grad_norm": 0.24799376726150513, |
| "learning_rate": 7.377915997506945e-06, |
| "loss": 0.4448, |
| "step": 272 |
| }, |
| { |
| "epoch": 1.2276810780460417, |
| "grad_norm": 0.29386043548583984, |
| "learning_rate": 7.354815225237468e-06, |
| "loss": 0.4617, |
| "step": 273 |
| }, |
| { |
| "epoch": 1.2321729365524985, |
| "grad_norm": 0.2598021924495697, |
| "learning_rate": 7.331649678785545e-06, |
| "loss": 0.4761, |
| "step": 274 |
| }, |
| { |
| "epoch": 1.2366647950589555, |
| "grad_norm": 0.2887316942214966, |
| "learning_rate": 7.308419995368616e-06, |
| "loss": 0.4835, |
| "step": 275 |
| }, |
| { |
| "epoch": 1.2411566535654126, |
| "grad_norm": 0.2421315759420395, |
| "learning_rate": 7.285126813968346e-06, |
| "loss": 0.4417, |
| "step": 276 |
| }, |
| { |
| "epoch": 1.2456485120718697, |
| "grad_norm": 0.296011358499527, |
| "learning_rate": 7.2617707753130465e-06, |
| "loss": 0.4778, |
| "step": 277 |
| }, |
| { |
| "epoch": 1.2501403705783267, |
| "grad_norm": 0.24542121589183807, |
| "learning_rate": 7.238352521860049e-06, |
| "loss": 0.4504, |
| "step": 278 |
| }, |
| { |
| "epoch": 1.2546322290847838, |
| "grad_norm": 0.2937444746494293, |
| "learning_rate": 7.214872697778037e-06, |
| "loss": 0.468, |
| "step": 279 |
| }, |
| { |
| "epoch": 1.2591240875912408, |
| "grad_norm": 0.2538372278213501, |
| "learning_rate": 7.191331948929323e-06, |
| "loss": 0.4629, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.263615946097698, |
| "grad_norm": 0.2601139545440674, |
| "learning_rate": 7.1677309228520865e-06, |
| "loss": 0.4675, |
| "step": 281 |
| }, |
| { |
| "epoch": 1.268107804604155, |
| "grad_norm": 0.254896342754364, |
| "learning_rate": 7.14407026874256e-06, |
| "loss": 0.4657, |
| "step": 282 |
| }, |
| { |
| "epoch": 1.272599663110612, |
| "grad_norm": 0.18791675567626953, |
| "learning_rate": 7.120350637437166e-06, |
| "loss": 0.4638, |
| "step": 283 |
| }, |
| { |
| "epoch": 1.277091521617069, |
| "grad_norm": 0.3108649253845215, |
| "learning_rate": 7.096572681394625e-06, |
| "loss": 0.4667, |
| "step": 284 |
| }, |
| { |
| "epoch": 1.2815833801235261, |
| "grad_norm": 0.21536403894424438, |
| "learning_rate": 7.072737054678004e-06, |
| "loss": 0.468, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.2860752386299832, |
| "grad_norm": 0.24769152700901031, |
| "learning_rate": 7.048844412936718e-06, |
| "loss": 0.4655, |
| "step": 286 |
| }, |
| { |
| "epoch": 1.2905670971364402, |
| "grad_norm": 0.2215883433818817, |
| "learning_rate": 7.024895413388508e-06, |
| "loss": 0.4816, |
| "step": 287 |
| }, |
| { |
| "epoch": 1.2950589556428973, |
| "grad_norm": 0.21110257506370544, |
| "learning_rate": 7.000890714801352e-06, |
| "loss": 0.4655, |
| "step": 288 |
| }, |
| { |
| "epoch": 1.2995508141493544, |
| "grad_norm": 0.21130134165287018, |
| "learning_rate": 6.976830977475346e-06, |
| "loss": 0.444, |
| "step": 289 |
| }, |
| { |
| "epoch": 1.3040426726558114, |
| "grad_norm": 0.2305009365081787, |
| "learning_rate": 6.952716863224551e-06, |
| "loss": 0.4665, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.3085345311622683, |
| "grad_norm": 0.24490021169185638, |
| "learning_rate": 6.928549035358772e-06, |
| "loss": 0.4675, |
| "step": 291 |
| }, |
| { |
| "epoch": 1.3130263896687255, |
| "grad_norm": 0.20088279247283936, |
| "learning_rate": 6.904328158665323e-06, |
| "loss": 0.4589, |
| "step": 292 |
| }, |
| { |
| "epoch": 1.3175182481751824, |
| "grad_norm": 0.25551822781562805, |
| "learning_rate": 6.880054899390744e-06, |
| "loss": 0.4707, |
| "step": 293 |
| }, |
| { |
| "epoch": 1.3220101066816397, |
| "grad_norm": 0.21504874527454376, |
| "learning_rate": 6.855729925222462e-06, |
| "loss": 0.4633, |
| "step": 294 |
| }, |
| { |
| "epoch": 1.3265019651880965, |
| "grad_norm": 0.2354370504617691, |
| "learning_rate": 6.831353905270433e-06, |
| "loss": 0.4739, |
| "step": 295 |
| }, |
| { |
| "epoch": 1.3309938236945535, |
| "grad_norm": 0.22273460030555725, |
| "learning_rate": 6.806927510048738e-06, |
| "loss": 0.4676, |
| "step": 296 |
| }, |
| { |
| "epoch": 1.3354856822010106, |
| "grad_norm": 0.20321142673492432, |
| "learning_rate": 6.782451411457137e-06, |
| "loss": 0.4581, |
| "step": 297 |
| }, |
| { |
| "epoch": 1.3399775407074677, |
| "grad_norm": 0.21091462671756744, |
| "learning_rate": 6.757926282762583e-06, |
| "loss": 0.4603, |
| "step": 298 |
| }, |
| { |
| "epoch": 1.3444693992139247, |
| "grad_norm": 0.22511275112628937, |
| "learning_rate": 6.733352798580708e-06, |
| "loss": 0.4584, |
| "step": 299 |
| }, |
| { |
| "epoch": 1.3489612577203818, |
| "grad_norm": 0.20873034000396729, |
| "learning_rate": 6.7087316348572626e-06, |
| "loss": 0.4768, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.3534531162268388, |
| "grad_norm": 0.19452911615371704, |
| "learning_rate": 6.684063468849528e-06, |
| "loss": 0.4687, |
| "step": 301 |
| }, |
| { |
| "epoch": 1.357944974733296, |
| "grad_norm": 0.2007768303155899, |
| "learning_rate": 6.659348979107679e-06, |
| "loss": 0.4602, |
| "step": 302 |
| }, |
| { |
| "epoch": 1.362436833239753, |
| "grad_norm": 0.18347479403018951, |
| "learning_rate": 6.634588845456122e-06, |
| "loss": 0.4565, |
| "step": 303 |
| }, |
| { |
| "epoch": 1.36692869174621, |
| "grad_norm": 0.2227877527475357, |
| "learning_rate": 6.609783748974802e-06, |
| "loss": 0.4554, |
| "step": 304 |
| }, |
| { |
| "epoch": 1.371420550252667, |
| "grad_norm": 0.2087058275938034, |
| "learning_rate": 6.584934371980452e-06, |
| "loss": 0.4552, |
| "step": 305 |
| }, |
| { |
| "epoch": 1.3759124087591241, |
| "grad_norm": 0.24951022863388062, |
| "learning_rate": 6.560041398007847e-06, |
| "loss": 0.46, |
| "step": 306 |
| }, |
| { |
| "epoch": 1.3804042672655812, |
| "grad_norm": 0.1927427351474762, |
| "learning_rate": 6.535105511790979e-06, |
| "loss": 0.4441, |
| "step": 307 |
| }, |
| { |
| "epoch": 1.3848961257720382, |
| "grad_norm": 0.2886252999305725, |
| "learning_rate": 6.510127399244235e-06, |
| "loss": 0.4769, |
| "step": 308 |
| }, |
| { |
| "epoch": 1.3893879842784953, |
| "grad_norm": 0.2261786013841629, |
| "learning_rate": 6.485107747443529e-06, |
| "loss": 0.4552, |
| "step": 309 |
| }, |
| { |
| "epoch": 1.3938798427849521, |
| "grad_norm": 0.22381217777729034, |
| "learning_rate": 6.460047244607397e-06, |
| "loss": 0.4576, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.3983717012914094, |
| "grad_norm": 0.21512459218502045, |
| "learning_rate": 6.434946580078072e-06, |
| "loss": 0.4417, |
| "step": 311 |
| }, |
| { |
| "epoch": 1.4028635597978663, |
| "grad_norm": 0.2732483446598053, |
| "learning_rate": 6.409806444302519e-06, |
| "loss": 0.4635, |
| "step": 312 |
| }, |
| { |
| "epoch": 1.4073554183043235, |
| "grad_norm": 0.2072446048259735, |
| "learning_rate": 6.384627528813439e-06, |
| "loss": 0.4544, |
| "step": 313 |
| }, |
| { |
| "epoch": 1.4118472768107804, |
| "grad_norm": 0.3143913447856903, |
| "learning_rate": 6.359410526210259e-06, |
| "loss": 0.49, |
| "step": 314 |
| }, |
| { |
| "epoch": 1.4163391353172374, |
| "grad_norm": 0.26985523104667664, |
| "learning_rate": 6.334156130140068e-06, |
| "loss": 0.4533, |
| "step": 315 |
| }, |
| { |
| "epoch": 1.4208309938236945, |
| "grad_norm": 0.25130948424339294, |
| "learning_rate": 6.308865035278539e-06, |
| "loss": 0.4536, |
| "step": 316 |
| }, |
| { |
| "epoch": 1.4253228523301515, |
| "grad_norm": 0.2829897403717041, |
| "learning_rate": 6.283537937310829e-06, |
| "loss": 0.484, |
| "step": 317 |
| }, |
| { |
| "epoch": 1.4298147108366086, |
| "grad_norm": 0.22410716116428375, |
| "learning_rate": 6.258175532912432e-06, |
| "loss": 0.4594, |
| "step": 318 |
| }, |
| { |
| "epoch": 1.4343065693430657, |
| "grad_norm": 0.2592792510986328, |
| "learning_rate": 6.232778519730024e-06, |
| "loss": 0.4709, |
| "step": 319 |
| }, |
| { |
| "epoch": 1.4387984278495227, |
| "grad_norm": 0.20447753369808197, |
| "learning_rate": 6.207347596362265e-06, |
| "loss": 0.4622, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.4432902863559798, |
| "grad_norm": 0.25977230072021484, |
| "learning_rate": 6.181883462340589e-06, |
| "loss": 0.4753, |
| "step": 321 |
| }, |
| { |
| "epoch": 1.4477821448624368, |
| "grad_norm": 0.20790189504623413, |
| "learning_rate": 6.1563868181099596e-06, |
| "loss": 0.4535, |
| "step": 322 |
| }, |
| { |
| "epoch": 1.452274003368894, |
| "grad_norm": 0.29562297463417053, |
| "learning_rate": 6.130858365009601e-06, |
| "loss": 0.4744, |
| "step": 323 |
| }, |
| { |
| "epoch": 1.456765861875351, |
| "grad_norm": 0.21025337278842926, |
| "learning_rate": 6.105298805253709e-06, |
| "loss": 0.4663, |
| "step": 324 |
| }, |
| { |
| "epoch": 1.461257720381808, |
| "grad_norm": 0.26838284730911255, |
| "learning_rate": 6.079708841912133e-06, |
| "loss": 0.4682, |
| "step": 325 |
| }, |
| { |
| "epoch": 1.465749578888265, |
| "grad_norm": 0.21380648016929626, |
| "learning_rate": 6.054089178891039e-06, |
| "loss": 0.4669, |
| "step": 326 |
| }, |
| { |
| "epoch": 1.4702414373947221, |
| "grad_norm": 0.2207087278366089, |
| "learning_rate": 6.028440520913545e-06, |
| "loss": 0.4414, |
| "step": 327 |
| }, |
| { |
| "epoch": 1.4747332959011792, |
| "grad_norm": 0.21855312585830688, |
| "learning_rate": 6.002763573500332e-06, |
| "loss": 0.4833, |
| "step": 328 |
| }, |
| { |
| "epoch": 1.479225154407636, |
| "grad_norm": 0.23827263712882996, |
| "learning_rate": 5.977059042950252e-06, |
| "loss": 0.4486, |
| "step": 329 |
| }, |
| { |
| "epoch": 1.4837170129140933, |
| "grad_norm": 0.22362874448299408, |
| "learning_rate": 5.951327636320878e-06, |
| "loss": 0.4761, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.4882088714205501, |
| "grad_norm": 0.23705798387527466, |
| "learning_rate": 5.925570061409077e-06, |
| "loss": 0.4667, |
| "step": 331 |
| }, |
| { |
| "epoch": 1.4927007299270074, |
| "grad_norm": 0.20592792332172394, |
| "learning_rate": 5.899787026731524e-06, |
| "loss": 0.4528, |
| "step": 332 |
| }, |
| { |
| "epoch": 1.4971925884334643, |
| "grad_norm": 0.1896170973777771, |
| "learning_rate": 5.873979241505219e-06, |
| "loss": 0.4618, |
| "step": 333 |
| }, |
| { |
| "epoch": 1.5016844469399215, |
| "grad_norm": 0.1971491128206253, |
| "learning_rate": 5.848147415627981e-06, |
| "loss": 0.4684, |
| "step": 334 |
| }, |
| { |
| "epoch": 1.5061763054463784, |
| "grad_norm": 0.21199488639831543, |
| "learning_rate": 5.822292259658914e-06, |
| "loss": 0.4653, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.5106681639528357, |
| "grad_norm": 0.19454307854175568, |
| "learning_rate": 5.79641448479887e-06, |
| "loss": 0.4739, |
| "step": 336 |
| }, |
| { |
| "epoch": 1.5151600224592925, |
| "grad_norm": 0.2176458239555359, |
| "learning_rate": 5.770514802870879e-06, |
| "loss": 0.4583, |
| "step": 337 |
| }, |
| { |
| "epoch": 1.5196518809657495, |
| "grad_norm": 0.21410858631134033, |
| "learning_rate": 5.744593926300573e-06, |
| "loss": 0.4494, |
| "step": 338 |
| }, |
| { |
| "epoch": 1.5241437394722066, |
| "grad_norm": 0.21628141403198242, |
| "learning_rate": 5.718652568096585e-06, |
| "loss": 0.4639, |
| "step": 339 |
| }, |
| { |
| "epoch": 1.5286355979786637, |
| "grad_norm": 0.2230045348405838, |
| "learning_rate": 5.6926914418309405e-06, |
| "loss": 0.4603, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.5331274564851207, |
| "grad_norm": 0.24906450510025024, |
| "learning_rate": 5.666711261619429e-06, |
| "loss": 0.4662, |
| "step": 341 |
| }, |
| { |
| "epoch": 1.5376193149915778, |
| "grad_norm": 0.19834677875041962, |
| "learning_rate": 5.640712742101954e-06, |
| "loss": 0.4582, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.5421111734980348, |
| "grad_norm": 0.2662658989429474, |
| "learning_rate": 5.614696598422885e-06, |
| "loss": 0.4748, |
| "step": 343 |
| }, |
| { |
| "epoch": 1.546603032004492, |
| "grad_norm": 0.22788119316101074, |
| "learning_rate": 5.5886635462113805e-06, |
| "loss": 0.4654, |
| "step": 344 |
| }, |
| { |
| "epoch": 1.551094890510949, |
| "grad_norm": 0.22741669416427612, |
| "learning_rate": 5.562614301561704e-06, |
| "loss": 0.4527, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.5555867490174058, |
| "grad_norm": 0.24191468954086304, |
| "learning_rate": 5.536549581013525e-06, |
| "loss": 0.4691, |
| "step": 346 |
| }, |
| { |
| "epoch": 1.560078607523863, |
| "grad_norm": 0.22361521422863007, |
| "learning_rate": 5.510470101532213e-06, |
| "loss": 0.4571, |
| "step": 347 |
| }, |
| { |
| "epoch": 1.56457046603032, |
| "grad_norm": 0.21503981947898865, |
| "learning_rate": 5.48437658048911e-06, |
| "loss": 0.447, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.5690623245367772, |
| "grad_norm": 0.22127224504947662, |
| "learning_rate": 5.4582697356418036e-06, |
| "loss": 0.4604, |
| "step": 349 |
| }, |
| { |
| "epoch": 1.573554183043234, |
| "grad_norm": 0.24467918276786804, |
| "learning_rate": 5.4321502851143785e-06, |
| "loss": 0.479, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.5780460415496913, |
| "grad_norm": 0.1886945217847824, |
| "learning_rate": 5.406018947377668e-06, |
| "loss": 0.4641, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.5825379000561481, |
| "grad_norm": 0.20987150073051453, |
| "learning_rate": 5.379876441229486e-06, |
| "loss": 0.4629, |
| "step": 352 |
| }, |
| { |
| "epoch": 1.5870297585626054, |
| "grad_norm": 0.19232037663459778, |
| "learning_rate": 5.353723485774859e-06, |
| "loss": 0.4367, |
| "step": 353 |
| }, |
| { |
| "epoch": 1.5915216170690623, |
| "grad_norm": 0.18378283083438873, |
| "learning_rate": 5.327560800406241e-06, |
| "loss": 0.48, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.5960134755755195, |
| "grad_norm": 0.22845953702926636, |
| "learning_rate": 5.301389104783731e-06, |
| "loss": 0.4695, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.6005053340819764, |
| "grad_norm": 0.1837320476770401, |
| "learning_rate": 5.275209118815274e-06, |
| "loss": 0.4402, |
| "step": 356 |
| }, |
| { |
| "epoch": 1.6049971925884334, |
| "grad_norm": 0.20027028024196625, |
| "learning_rate": 5.249021562636857e-06, |
| "loss": 0.4538, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.6094890510948905, |
| "grad_norm": 0.2599356472492218, |
| "learning_rate": 5.222827156592701e-06, |
| "loss": 0.4719, |
| "step": 358 |
| }, |
| { |
| "epoch": 1.6139809096013475, |
| "grad_norm": 0.20169083774089813, |
| "learning_rate": 5.196626621215449e-06, |
| "loss": 0.4722, |
| "step": 359 |
| }, |
| { |
| "epoch": 1.6184727681078046, |
| "grad_norm": 0.2602737247943878, |
| "learning_rate": 5.170420677206344e-06, |
| "loss": 0.4567, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.6229646266142617, |
| "grad_norm": 0.20139826834201813, |
| "learning_rate": 5.144210045415402e-06, |
| "loss": 0.4607, |
| "step": 361 |
| }, |
| { |
| "epoch": 1.6274564851207187, |
| "grad_norm": 0.19559621810913086, |
| "learning_rate": 5.117995446821592e-06, |
| "loss": 0.4652, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.6319483436271758, |
| "grad_norm": 0.21619027853012085, |
| "learning_rate": 5.091777602512993e-06, |
| "loss": 0.4381, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.6364402021336328, |
| "grad_norm": 0.20571349561214447, |
| "learning_rate": 5.065557233666968e-06, |
| "loss": 0.4665, |
| "step": 364 |
| }, |
| { |
| "epoch": 1.6409320606400897, |
| "grad_norm": 0.2172069400548935, |
| "learning_rate": 5.039335061530319e-06, |
| "loss": 0.4463, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.645423919146547, |
| "grad_norm": 0.20271949470043182, |
| "learning_rate": 5.013111807399455e-06, |
| "loss": 0.458, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.6499157776530038, |
| "grad_norm": 0.19836212694644928, |
| "learning_rate": 4.986888192600546e-06, |
| "loss": 0.461, |
| "step": 367 |
| }, |
| { |
| "epoch": 1.654407636159461, |
| "grad_norm": 0.1832582950592041, |
| "learning_rate": 4.960664938469683e-06, |
| "loss": 0.4393, |
| "step": 368 |
| }, |
| { |
| "epoch": 1.658899494665918, |
| "grad_norm": 0.23569650948047638, |
| "learning_rate": 4.934442766333034e-06, |
| "loss": 0.4717, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.6633913531723752, |
| "grad_norm": 0.186725914478302, |
| "learning_rate": 4.908222397487009e-06, |
| "loss": 0.4522, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.667883211678832, |
| "grad_norm": 0.21094153821468353, |
| "learning_rate": 4.88200455317841e-06, |
| "loss": 0.4543, |
| "step": 371 |
| }, |
| { |
| "epoch": 1.6723750701852893, |
| "grad_norm": 0.24533981084823608, |
| "learning_rate": 4.8557899545846e-06, |
| "loss": 0.4714, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.6768669286917461, |
| "grad_norm": 0.19219504296779633, |
| "learning_rate": 4.829579322793659e-06, |
| "loss": 0.4756, |
| "step": 373 |
| }, |
| { |
| "epoch": 1.6813587871982034, |
| "grad_norm": 0.22109220921993256, |
| "learning_rate": 4.8033733787845535e-06, |
| "loss": 0.4423, |
| "step": 374 |
| }, |
| { |
| "epoch": 1.6858506457046603, |
| "grad_norm": 0.21374572813510895, |
| "learning_rate": 4.7771728434073005e-06, |
| "loss": 0.4526, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.6903425042111173, |
| "grad_norm": 0.21615076065063477, |
| "learning_rate": 4.7509784373631446e-06, |
| "loss": 0.4657, |
| "step": 376 |
| }, |
| { |
| "epoch": 1.6948343627175744, |
| "grad_norm": 0.2250211238861084, |
| "learning_rate": 4.724790881184727e-06, |
| "loss": 0.4621, |
| "step": 377 |
| }, |
| { |
| "epoch": 1.6993262212240314, |
| "grad_norm": 0.21998076140880585, |
| "learning_rate": 4.69861089521627e-06, |
| "loss": 0.4613, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.7038180797304885, |
| "grad_norm": 0.20503737032413483, |
| "learning_rate": 4.672439199593761e-06, |
| "loss": 0.4529, |
| "step": 379 |
| }, |
| { |
| "epoch": 1.7083099382369455, |
| "grad_norm": 0.2038574367761612, |
| "learning_rate": 4.646276514225143e-06, |
| "loss": 0.4556, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.7128017967434026, |
| "grad_norm": 0.18167847394943237, |
| "learning_rate": 4.6201235587705155e-06, |
| "loss": 0.4734, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.7172936552498597, |
| "grad_norm": 0.1952500194311142, |
| "learning_rate": 4.593981052622334e-06, |
| "loss": 0.4542, |
| "step": 382 |
| }, |
| { |
| "epoch": 1.7217855137563167, |
| "grad_norm": 0.2038043886423111, |
| "learning_rate": 4.567849714885622e-06, |
| "loss": 0.482, |
| "step": 383 |
| }, |
| { |
| "epoch": 1.7262773722627736, |
| "grad_norm": 0.1942799836397171, |
| "learning_rate": 4.541730264358198e-06, |
| "loss": 0.4502, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.7307692307692308, |
| "grad_norm": 0.17968258261680603, |
| "learning_rate": 4.515623419510891e-06, |
| "loss": 0.4595, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.7352610892756877, |
| "grad_norm": 0.21333391964435577, |
| "learning_rate": 4.489529898467789e-06, |
| "loss": 0.4742, |
| "step": 386 |
| }, |
| { |
| "epoch": 1.739752947782145, |
| "grad_norm": 0.19606302678585052, |
| "learning_rate": 4.463450418986477e-06, |
| "loss": 0.4581, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.7442448062886018, |
| "grad_norm": 0.17203934490680695, |
| "learning_rate": 4.4373856984382985e-06, |
| "loss": 0.449, |
| "step": 388 |
| }, |
| { |
| "epoch": 1.748736664795059, |
| "grad_norm": 0.19502601027488708, |
| "learning_rate": 4.411336453788622e-06, |
| "loss": 0.472, |
| "step": 389 |
| }, |
| { |
| "epoch": 1.753228523301516, |
| "grad_norm": 0.19490373134613037, |
| "learning_rate": 4.3853034015771176e-06, |
| "loss": 0.4509, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.7577203818079732, |
| "grad_norm": 0.18702197074890137, |
| "learning_rate": 4.3592872578980495e-06, |
| "loss": 0.4867, |
| "step": 391 |
| }, |
| { |
| "epoch": 1.76221224031443, |
| "grad_norm": 0.185566708445549, |
| "learning_rate": 4.333288738380574e-06, |
| "loss": 0.4488, |
| "step": 392 |
| }, |
| { |
| "epoch": 1.7667040988208873, |
| "grad_norm": 0.17845405638217926, |
| "learning_rate": 4.30730855816906e-06, |
| "loss": 0.4657, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.7711959573273441, |
| "grad_norm": 0.17764534056186676, |
| "learning_rate": 4.281347431903416e-06, |
| "loss": 0.4499, |
| "step": 394 |
| }, |
| { |
| "epoch": 1.7756878158338012, |
| "grad_norm": 0.19315990805625916, |
| "learning_rate": 4.255406073699428e-06, |
| "loss": 0.4497, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.7801796743402583, |
| "grad_norm": 0.1673203557729721, |
| "learning_rate": 4.229485197129122e-06, |
| "loss": 0.4462, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.7846715328467153, |
| "grad_norm": 0.18644391000270844, |
| "learning_rate": 4.203585515201131e-06, |
| "loss": 0.4469, |
| "step": 397 |
| }, |
| { |
| "epoch": 1.7891633913531724, |
| "grad_norm": 0.17737650871276855, |
| "learning_rate": 4.177707740341088e-06, |
| "loss": 0.4605, |
| "step": 398 |
| }, |
| { |
| "epoch": 1.7936552498596294, |
| "grad_norm": 0.17971952259540558, |
| "learning_rate": 4.151852584372021e-06, |
| "loss": 0.4473, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.7981471083660865, |
| "grad_norm": 0.1814514845609665, |
| "learning_rate": 4.1260207584947825e-06, |
| "loss": 0.466, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.8026389668725435, |
| "grad_norm": 0.17787028849124908, |
| "learning_rate": 4.100212973268478e-06, |
| "loss": 0.4656, |
| "step": 401 |
| }, |
| { |
| "epoch": 1.8071308253790006, |
| "grad_norm": 0.2006688117980957, |
| "learning_rate": 4.074429938590924e-06, |
| "loss": 0.4484, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.8116226838854577, |
| "grad_norm": 0.20157967507839203, |
| "learning_rate": 4.048672363679124e-06, |
| "loss": 0.4735, |
| "step": 403 |
| }, |
| { |
| "epoch": 1.8161145423919147, |
| "grad_norm": 0.21290093660354614, |
| "learning_rate": 4.022940957049752e-06, |
| "loss": 0.4705, |
| "step": 404 |
| }, |
| { |
| "epoch": 1.8206064008983716, |
| "grad_norm": 0.18965761363506317, |
| "learning_rate": 3.99723642649967e-06, |
| "loss": 0.4534, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.8250982594048288, |
| "grad_norm": 0.19589713215827942, |
| "learning_rate": 3.971559479086459e-06, |
| "loss": 0.4587, |
| "step": 406 |
| }, |
| { |
| "epoch": 1.8295901179112857, |
| "grad_norm": 0.16529035568237305, |
| "learning_rate": 3.945910821108963e-06, |
| "loss": 0.4587, |
| "step": 407 |
| }, |
| { |
| "epoch": 1.834081976417743, |
| "grad_norm": 0.17363354563713074, |
| "learning_rate": 3.9202911580878685e-06, |
| "loss": 0.4409, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.8385738349241998, |
| "grad_norm": 0.1951248049736023, |
| "learning_rate": 3.894701194746291e-06, |
| "loss": 0.4534, |
| "step": 409 |
| }, |
| { |
| "epoch": 1.843065693430657, |
| "grad_norm": 0.18388287723064423, |
| "learning_rate": 3.869141634990399e-06, |
| "loss": 0.474, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.847557551937114, |
| "grad_norm": 0.17535457015037537, |
| "learning_rate": 3.843613181890042e-06, |
| "loss": 0.4469, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.8520494104435712, |
| "grad_norm": 0.1836119443178177, |
| "learning_rate": 3.818116537659412e-06, |
| "loss": 0.4536, |
| "step": 412 |
| }, |
| { |
| "epoch": 1.856541268950028, |
| "grad_norm": 0.18821248412132263, |
| "learning_rate": 3.7926524036377366e-06, |
| "loss": 0.477, |
| "step": 413 |
| }, |
| { |
| "epoch": 1.861033127456485, |
| "grad_norm": 0.19100096821784973, |
| "learning_rate": 3.767221480269978e-06, |
| "loss": 0.4336, |
| "step": 414 |
| }, |
| { |
| "epoch": 1.8655249859629421, |
| "grad_norm": 0.1809721142053604, |
| "learning_rate": 3.741824467087569e-06, |
| "loss": 0.4539, |
| "step": 415 |
| }, |
| { |
| "epoch": 1.8700168444693992, |
| "grad_norm": 0.22617210447788239, |
| "learning_rate": 3.7164620626891724e-06, |
| "loss": 0.4883, |
| "step": 416 |
| }, |
| { |
| "epoch": 1.8745087029758563, |
| "grad_norm": 0.1782565861940384, |
| "learning_rate": 3.6911349647214623e-06, |
| "loss": 0.4498, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.8790005614823133, |
| "grad_norm": 0.16668500006198883, |
| "learning_rate": 3.665843869859934e-06, |
| "loss": 0.4563, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.8834924199887704, |
| "grad_norm": 0.2230617105960846, |
| "learning_rate": 3.640589473789742e-06, |
| "loss": 0.4695, |
| "step": 419 |
| }, |
| { |
| "epoch": 1.8879842784952274, |
| "grad_norm": 0.17857353389263153, |
| "learning_rate": 3.6153724711865623e-06, |
| "loss": 0.4445, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.8924761370016845, |
| "grad_norm": 0.19203616678714752, |
| "learning_rate": 3.5901935556974837e-06, |
| "loss": 0.4559, |
| "step": 421 |
| }, |
| { |
| "epoch": 1.8969679955081415, |
| "grad_norm": 0.18788589537143707, |
| "learning_rate": 3.56505341992193e-06, |
| "loss": 0.4543, |
| "step": 422 |
| }, |
| { |
| "epoch": 1.9014598540145986, |
| "grad_norm": 0.17305254936218262, |
| "learning_rate": 3.539952755392605e-06, |
| "loss": 0.4567, |
| "step": 423 |
| }, |
| { |
| "epoch": 1.9059517125210554, |
| "grad_norm": 0.21953020989894867, |
| "learning_rate": 3.514892252556474e-06, |
| "loss": 0.4717, |
| "step": 424 |
| }, |
| { |
| "epoch": 1.9104435710275127, |
| "grad_norm": 0.15481020510196686, |
| "learning_rate": 3.4898726007557655e-06, |
| "loss": 0.4424, |
| "step": 425 |
| }, |
| { |
| "epoch": 1.9149354295339696, |
| "grad_norm": 0.18283706903457642, |
| "learning_rate": 3.464894488209022e-06, |
| "loss": 0.4467, |
| "step": 426 |
| }, |
| { |
| "epoch": 1.9194272880404268, |
| "grad_norm": 0.18225404620170593, |
| "learning_rate": 3.439958601992153e-06, |
| "loss": 0.4805, |
| "step": 427 |
| }, |
| { |
| "epoch": 1.9239191465468837, |
| "grad_norm": 0.15692797303199768, |
| "learning_rate": 3.415065628019547e-06, |
| "loss": 0.4536, |
| "step": 428 |
| }, |
| { |
| "epoch": 1.928411005053341, |
| "grad_norm": 0.17508453130722046, |
| "learning_rate": 3.3902162510252e-06, |
| "loss": 0.4567, |
| "step": 429 |
| }, |
| { |
| "epoch": 1.9329028635597978, |
| "grad_norm": 0.1812361180782318, |
| "learning_rate": 3.365411154543878e-06, |
| "loss": 0.4466, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.937394722066255, |
| "grad_norm": 0.17098145186901093, |
| "learning_rate": 3.3406510208923228e-06, |
| "loss": 0.4563, |
| "step": 431 |
| }, |
| { |
| "epoch": 1.941886580572712, |
| "grad_norm": 0.19324849545955658, |
| "learning_rate": 3.3159365311504732e-06, |
| "loss": 0.4649, |
| "step": 432 |
| }, |
| { |
| "epoch": 1.946378439079169, |
| "grad_norm": 0.17696388065814972, |
| "learning_rate": 3.291268365142738e-06, |
| "loss": 0.4518, |
| "step": 433 |
| }, |
| { |
| "epoch": 1.950870297585626, |
| "grad_norm": 0.1926283836364746, |
| "learning_rate": 3.2666472014192942e-06, |
| "loss": 0.4457, |
| "step": 434 |
| }, |
| { |
| "epoch": 1.955362156092083, |
| "grad_norm": 0.18097148835659027, |
| "learning_rate": 3.2420737172374184e-06, |
| "loss": 0.4649, |
| "step": 435 |
| }, |
| { |
| "epoch": 1.9598540145985401, |
| "grad_norm": 0.2086239606142044, |
| "learning_rate": 3.217548588542864e-06, |
| "loss": 0.466, |
| "step": 436 |
| }, |
| { |
| "epoch": 1.9643458731049972, |
| "grad_norm": 0.17721500992774963, |
| "learning_rate": 3.1930724899512633e-06, |
| "loss": 0.4576, |
| "step": 437 |
| }, |
| { |
| "epoch": 1.9688377316114543, |
| "grad_norm": 0.18082091212272644, |
| "learning_rate": 3.1686460947295695e-06, |
| "loss": 0.4543, |
| "step": 438 |
| }, |
| { |
| "epoch": 1.9733295901179113, |
| "grad_norm": 0.19457915425300598, |
| "learning_rate": 3.1442700747775413e-06, |
| "loss": 0.4489, |
| "step": 439 |
| }, |
| { |
| "epoch": 1.9778214486243684, |
| "grad_norm": 0.18913620710372925, |
| "learning_rate": 3.1199451006092586e-06, |
| "loss": 0.4631, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.9823133071308254, |
| "grad_norm": 0.18790724873542786, |
| "learning_rate": 3.0956718413346785e-06, |
| "loss": 0.4499, |
| "step": 441 |
| }, |
| { |
| "epoch": 1.9868051656372825, |
| "grad_norm": 0.16851408779621124, |
| "learning_rate": 3.0714509646412295e-06, |
| "loss": 0.4436, |
| "step": 442 |
| }, |
| { |
| "epoch": 1.9912970241437393, |
| "grad_norm": 0.19248563051223755, |
| "learning_rate": 3.0472831367754496e-06, |
| "loss": 0.4699, |
| "step": 443 |
| }, |
| { |
| "epoch": 1.9957888826501966, |
| "grad_norm": 0.17517177760601044, |
| "learning_rate": 3.0231690225246537e-06, |
| "loss": 0.454, |
| "step": 444 |
| }, |
| { |
| "epoch": 2.0016844469399215, |
| "grad_norm": 0.30906590819358826, |
| "learning_rate": 2.999109285198649e-06, |
| "loss": 0.7469, |
| "step": 445 |
| }, |
| { |
| "epoch": 2.0061763054463784, |
| "grad_norm": 0.18735533952713013, |
| "learning_rate": 2.9751045866114926e-06, |
| "loss": 0.45, |
| "step": 446 |
| }, |
| { |
| "epoch": 2.0106681639528357, |
| "grad_norm": 0.17937350273132324, |
| "learning_rate": 2.9511555870632824e-06, |
| "loss": 0.4468, |
| "step": 447 |
| }, |
| { |
| "epoch": 2.0151600224592925, |
| "grad_norm": 0.18137559294700623, |
| "learning_rate": 2.927262945321998e-06, |
| "loss": 0.4341, |
| "step": 448 |
| }, |
| { |
| "epoch": 2.0196518809657498, |
| "grad_norm": 0.18004953861236572, |
| "learning_rate": 2.903427318605376e-06, |
| "loss": 0.4426, |
| "step": 449 |
| }, |
| { |
| "epoch": 2.0241437394722066, |
| "grad_norm": 0.17482782900333405, |
| "learning_rate": 2.8796493625628357e-06, |
| "loss": 0.4584, |
| "step": 450 |
| }, |
| { |
| "epoch": 2.028635597978664, |
| "grad_norm": 0.20240633189678192, |
| "learning_rate": 2.8559297312574417e-06, |
| "loss": 0.4589, |
| "step": 451 |
| }, |
| { |
| "epoch": 2.0331274564851207, |
| "grad_norm": 0.15797480940818787, |
| "learning_rate": 2.8322690771479135e-06, |
| "loss": 0.4416, |
| "step": 452 |
| }, |
| { |
| "epoch": 2.0376193149915776, |
| "grad_norm": 0.16248145699501038, |
| "learning_rate": 2.808668051070678e-06, |
| "loss": 0.4356, |
| "step": 453 |
| }, |
| { |
| "epoch": 2.042111173498035, |
| "grad_norm": 0.17543736100196838, |
| "learning_rate": 2.7851273022219645e-06, |
| "loss": 0.4531, |
| "step": 454 |
| }, |
| { |
| "epoch": 2.0466030320044917, |
| "grad_norm": 0.18104536831378937, |
| "learning_rate": 2.7616474781399527e-06, |
| "loss": 0.4566, |
| "step": 455 |
| }, |
| { |
| "epoch": 2.051094890510949, |
| "grad_norm": 0.1631019413471222, |
| "learning_rate": 2.7382292246869548e-06, |
| "loss": 0.4345, |
| "step": 456 |
| }, |
| { |
| "epoch": 2.055586749017406, |
| "grad_norm": 0.16401013731956482, |
| "learning_rate": 2.7148731860316544e-06, |
| "loss": 0.4466, |
| "step": 457 |
| }, |
| { |
| "epoch": 2.060078607523863, |
| "grad_norm": 0.18238060176372528, |
| "learning_rate": 2.6915800046313852e-06, |
| "loss": 0.4587, |
| "step": 458 |
| }, |
| { |
| "epoch": 2.06457046603032, |
| "grad_norm": 0.1778591126203537, |
| "learning_rate": 2.6683503212144563e-06, |
| "loss": 0.4652, |
| "step": 459 |
| }, |
| { |
| "epoch": 2.069062324536777, |
| "grad_norm": 0.16754575073719025, |
| "learning_rate": 2.645184774762533e-06, |
| "loss": 0.4265, |
| "step": 460 |
| }, |
| { |
| "epoch": 2.073554183043234, |
| "grad_norm": 0.16498114168643951, |
| "learning_rate": 2.6220840024930562e-06, |
| "loss": 0.4335, |
| "step": 461 |
| }, |
| { |
| "epoch": 2.0780460415496913, |
| "grad_norm": 0.18743063509464264, |
| "learning_rate": 2.599048639841717e-06, |
| "loss": 0.4472, |
| "step": 462 |
| }, |
| { |
| "epoch": 2.082537900056148, |
| "grad_norm": 0.1664198338985443, |
| "learning_rate": 2.5760793204449737e-06, |
| "loss": 0.4468, |
| "step": 463 |
| }, |
| { |
| "epoch": 2.0870297585626054, |
| "grad_norm": 0.16629664599895477, |
| "learning_rate": 2.553176676122627e-06, |
| "loss": 0.4449, |
| "step": 464 |
| }, |
| { |
| "epoch": 2.0915216170690623, |
| "grad_norm": 0.1749490648508072, |
| "learning_rate": 2.530341336860439e-06, |
| "loss": 0.4378, |
| "step": 465 |
| }, |
| { |
| "epoch": 2.0960134755755195, |
| "grad_norm": 0.15219521522521973, |
| "learning_rate": 2.5075739307928017e-06, |
| "loss": 0.4317, |
| "step": 466 |
| }, |
| { |
| "epoch": 2.1005053340819764, |
| "grad_norm": 0.16345295310020447, |
| "learning_rate": 2.484875084185462e-06, |
| "loss": 0.4491, |
| "step": 467 |
| }, |
| { |
| "epoch": 2.1049971925884337, |
| "grad_norm": 0.162262961268425, |
| "learning_rate": 2.462245421418292e-06, |
| "loss": 0.4477, |
| "step": 468 |
| }, |
| { |
| "epoch": 2.1094890510948905, |
| "grad_norm": 0.16510337591171265, |
| "learning_rate": 2.4396855649681166e-06, |
| "loss": 0.4344, |
| "step": 469 |
| }, |
| { |
| "epoch": 2.1139809096013478, |
| "grad_norm": 0.17441292107105255, |
| "learning_rate": 2.4171961353915914e-06, |
| "loss": 0.4657, |
| "step": 470 |
| }, |
| { |
| "epoch": 2.1184727681078046, |
| "grad_norm": 0.15476343035697937, |
| "learning_rate": 2.394777751308129e-06, |
| "loss": 0.4377, |
| "step": 471 |
| }, |
| { |
| "epoch": 2.1229646266142614, |
| "grad_norm": 0.16448856890201569, |
| "learning_rate": 2.372431029382888e-06, |
| "loss": 0.4497, |
| "step": 472 |
| }, |
| { |
| "epoch": 2.1274564851207187, |
| "grad_norm": 0.18509458005428314, |
| "learning_rate": 2.350156584309804e-06, |
| "loss": 0.4511, |
| "step": 473 |
| }, |
| { |
| "epoch": 2.1319483436271756, |
| "grad_norm": 0.1604773998260498, |
| "learning_rate": 2.3279550287946883e-06, |
| "loss": 0.4582, |
| "step": 474 |
| }, |
| { |
| "epoch": 2.136440202133633, |
| "grad_norm": 0.17485181987285614, |
| "learning_rate": 2.305826973538366e-06, |
| "loss": 0.4403, |
| "step": 475 |
| }, |
| { |
| "epoch": 2.1409320606400897, |
| "grad_norm": 0.1811332404613495, |
| "learning_rate": 2.2837730272198886e-06, |
| "loss": 0.4436, |
| "step": 476 |
| }, |
| { |
| "epoch": 2.145423919146547, |
| "grad_norm": 0.15909186005592346, |
| "learning_rate": 2.2617937964797786e-06, |
| "loss": 0.4432, |
| "step": 477 |
| }, |
| { |
| "epoch": 2.149915777653004, |
| "grad_norm": 0.1537752002477646, |
| "learning_rate": 2.2398898859033496e-06, |
| "loss": 0.4398, |
| "step": 478 |
| }, |
| { |
| "epoch": 2.154407636159461, |
| "grad_norm": 0.14990779757499695, |
| "learning_rate": 2.2180618980040747e-06, |
| "loss": 0.4434, |
| "step": 479 |
| }, |
| { |
| "epoch": 2.158899494665918, |
| "grad_norm": 0.15770702064037323, |
| "learning_rate": 2.196310433207013e-06, |
| "loss": 0.445, |
| "step": 480 |
| }, |
| { |
| "epoch": 2.163391353172375, |
| "grad_norm": 0.1663922518491745, |
| "learning_rate": 2.174636089832293e-06, |
| "loss": 0.4624, |
| "step": 481 |
| }, |
| { |
| "epoch": 2.167883211678832, |
| "grad_norm": 0.1494603157043457, |
| "learning_rate": 2.1530394640786567e-06, |
| "loss": 0.4095, |
| "step": 482 |
| }, |
| { |
| "epoch": 2.1723750701852893, |
| "grad_norm": 0.16668330132961273, |
| "learning_rate": 2.131521150007056e-06, |
| "loss": 0.4749, |
| "step": 483 |
| }, |
| { |
| "epoch": 2.176866928691746, |
| "grad_norm": 0.156080424785614, |
| "learning_rate": 2.110081739524316e-06, |
| "loss": 0.4377, |
| "step": 484 |
| }, |
| { |
| "epoch": 2.1813587871982034, |
| "grad_norm": 0.1667332947254181, |
| "learning_rate": 2.0887218223668493e-06, |
| "loss": 0.4634, |
| "step": 485 |
| }, |
| { |
| "epoch": 2.1858506457046603, |
| "grad_norm": 0.16652284562587738, |
| "learning_rate": 2.0674419860844385e-06, |
| "loss": 0.4635, |
| "step": 486 |
| }, |
| { |
| "epoch": 2.1903425042111175, |
| "grad_norm": 0.17611587047576904, |
| "learning_rate": 2.046242816024071e-06, |
| "loss": 0.4333, |
| "step": 487 |
| }, |
| { |
| "epoch": 2.1948343627175744, |
| "grad_norm": 0.170632004737854, |
| "learning_rate": 2.0251248953138377e-06, |
| "loss": 0.4577, |
| "step": 488 |
| }, |
| { |
| "epoch": 2.199326221224031, |
| "grad_norm": 0.1769552379846573, |
| "learning_rate": 2.0040888048468954e-06, |
| "loss": 0.4404, |
| "step": 489 |
| }, |
| { |
| "epoch": 2.2038180797304885, |
| "grad_norm": 0.16041500866413116, |
| "learning_rate": 1.9831351232654874e-06, |
| "loss": 0.4354, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.2083099382369458, |
| "grad_norm": 0.15111730992794037, |
| "learning_rate": 1.962264426945023e-06, |
| "loss": 0.4412, |
| "step": 491 |
| }, |
| { |
| "epoch": 2.2128017967434026, |
| "grad_norm": 0.15624019503593445, |
| "learning_rate": 1.9414772899782274e-06, |
| "loss": 0.438, |
| "step": 492 |
| }, |
| { |
| "epoch": 2.2172936552498594, |
| "grad_norm": 0.1754332333803177, |
| "learning_rate": 1.920774284159353e-06, |
| "loss": 0.4604, |
| "step": 493 |
| }, |
| { |
| "epoch": 2.2217855137563167, |
| "grad_norm": 0.16165252029895782, |
| "learning_rate": 1.9001559789684403e-06, |
| "loss": 0.4283, |
| "step": 494 |
| }, |
| { |
| "epoch": 2.2262773722627736, |
| "grad_norm": 0.15245720744132996, |
| "learning_rate": 1.8796229415556628e-06, |
| "loss": 0.4331, |
| "step": 495 |
| }, |
| { |
| "epoch": 2.230769230769231, |
| "grad_norm": 0.1491118222475052, |
| "learning_rate": 1.859175736725724e-06, |
| "loss": 0.4371, |
| "step": 496 |
| }, |
| { |
| "epoch": 2.2352610892756877, |
| "grad_norm": 0.1607341170310974, |
| "learning_rate": 1.8388149269223153e-06, |
| "loss": 0.4443, |
| "step": 497 |
| }, |
| { |
| "epoch": 2.239752947782145, |
| "grad_norm": 0.16271331906318665, |
| "learning_rate": 1.8185410722126556e-06, |
| "loss": 0.4367, |
| "step": 498 |
| }, |
| { |
| "epoch": 2.244244806288602, |
| "grad_norm": 0.15422794222831726, |
| "learning_rate": 1.7983547302720773e-06, |
| "loss": 0.4297, |
| "step": 499 |
| }, |
| { |
| "epoch": 2.248736664795059, |
| "grad_norm": 0.16086190938949585, |
| "learning_rate": 1.7782564563686882e-06, |
| "loss": 0.4595, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.253228523301516, |
| "grad_norm": 0.1475391536951065, |
| "learning_rate": 1.7582468033480992e-06, |
| "loss": 0.4418, |
| "step": 501 |
| }, |
| { |
| "epoch": 2.257720381807973, |
| "grad_norm": 0.16517473757266998, |
| "learning_rate": 1.7383263216182155e-06, |
| "loss": 0.4611, |
| "step": 502 |
| }, |
| { |
| "epoch": 2.26221224031443, |
| "grad_norm": 0.14682336151599884, |
| "learning_rate": 1.7184955591340974e-06, |
| "loss": 0.4302, |
| "step": 503 |
| }, |
| { |
| "epoch": 2.2667040988208873, |
| "grad_norm": 0.16191935539245605, |
| "learning_rate": 1.6987550613828863e-06, |
| "loss": 0.4728, |
| "step": 504 |
| }, |
| { |
| "epoch": 2.271195957327344, |
| "grad_norm": 0.1535215973854065, |
| "learning_rate": 1.6791053713688022e-06, |
| "loss": 0.4396, |
| "step": 505 |
| }, |
| { |
| "epoch": 2.2756878158338014, |
| "grad_norm": 0.1514642834663391, |
| "learning_rate": 1.6595470295982047e-06, |
| "loss": 0.4341, |
| "step": 506 |
| }, |
| { |
| "epoch": 2.2801796743402583, |
| "grad_norm": 0.1438160538673401, |
| "learning_rate": 1.6400805740647269e-06, |
| "loss": 0.4426, |
| "step": 507 |
| }, |
| { |
| "epoch": 2.2846715328467155, |
| "grad_norm": 0.1464770883321762, |
| "learning_rate": 1.6207065402344746e-06, |
| "loss": 0.4757, |
| "step": 508 |
| }, |
| { |
| "epoch": 2.2891633913531724, |
| "grad_norm": 0.14882051944732666, |
| "learning_rate": 1.6014254610313035e-06, |
| "loss": 0.4402, |
| "step": 509 |
| }, |
| { |
| "epoch": 2.293655249859629, |
| "grad_norm": 0.14686354994773865, |
| "learning_rate": 1.5822378668221511e-06, |
| "loss": 0.4506, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.2981471083660865, |
| "grad_norm": 0.16808481514453888, |
| "learning_rate": 1.563144285402453e-06, |
| "loss": 0.4535, |
| "step": 511 |
| }, |
| { |
| "epoch": 2.3026389668725433, |
| "grad_norm": 0.1671183556318283, |
| "learning_rate": 1.5441452419816238e-06, |
| "loss": 0.4576, |
| "step": 512 |
| }, |
| { |
| "epoch": 2.3071308253790006, |
| "grad_norm": 0.158988356590271, |
| "learning_rate": 1.5252412591686105e-06, |
| "loss": 0.4606, |
| "step": 513 |
| }, |
| { |
| "epoch": 2.3116226838854574, |
| "grad_norm": 0.1585473269224167, |
| "learning_rate": 1.5064328569575166e-06, |
| "loss": 0.4364, |
| "step": 514 |
| }, |
| { |
| "epoch": 2.3161145423919147, |
| "grad_norm": 0.1624506711959839, |
| "learning_rate": 1.4877205527132983e-06, |
| "loss": 0.4452, |
| "step": 515 |
| }, |
| { |
| "epoch": 2.3206064008983716, |
| "grad_norm": 0.14651590585708618, |
| "learning_rate": 1.4691048611575337e-06, |
| "loss": 0.432, |
| "step": 516 |
| }, |
| { |
| "epoch": 2.325098259404829, |
| "grad_norm": 0.14389865100383759, |
| "learning_rate": 1.4505862943542643e-06, |
| "loss": 0.4517, |
| "step": 517 |
| }, |
| { |
| "epoch": 2.3295901179112857, |
| "grad_norm": 0.15029338002204895, |
| "learning_rate": 1.4321653616959096e-06, |
| "loss": 0.4548, |
| "step": 518 |
| }, |
| { |
| "epoch": 2.334081976417743, |
| "grad_norm": 0.14938291907310486, |
| "learning_rate": 1.4138425698892555e-06, |
| "loss": 0.444, |
| "step": 519 |
| }, |
| { |
| "epoch": 2.3385738349242, |
| "grad_norm": 0.15582537651062012, |
| "learning_rate": 1.3956184229415148e-06, |
| "loss": 0.4343, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.343065693430657, |
| "grad_norm": 0.16049213707447052, |
| "learning_rate": 1.3774934221464643e-06, |
| "loss": 0.4866, |
| "step": 521 |
| }, |
| { |
| "epoch": 2.347557551937114, |
| "grad_norm": 0.1587788462638855, |
| "learning_rate": 1.3594680660706572e-06, |
| "loss": 0.4259, |
| "step": 522 |
| }, |
| { |
| "epoch": 2.352049410443571, |
| "grad_norm": 0.15312546491622925, |
| "learning_rate": 1.341542850539706e-06, |
| "loss": 0.4567, |
| "step": 523 |
| }, |
| { |
| "epoch": 2.356541268950028, |
| "grad_norm": 0.16400395333766937, |
| "learning_rate": 1.323718268624647e-06, |
| "loss": 0.437, |
| "step": 524 |
| }, |
| { |
| "epoch": 2.3610331274564853, |
| "grad_norm": 0.15856412053108215, |
| "learning_rate": 1.3059948106283725e-06, |
| "loss": 0.4568, |
| "step": 525 |
| }, |
| { |
| "epoch": 2.365524985962942, |
| "grad_norm": 0.16507361829280853, |
| "learning_rate": 1.2883729640721532e-06, |
| "loss": 0.4369, |
| "step": 526 |
| }, |
| { |
| "epoch": 2.370016844469399, |
| "grad_norm": 0.1538458913564682, |
| "learning_rate": 1.2708532136822156e-06, |
| "loss": 0.4352, |
| "step": 527 |
| }, |
| { |
| "epoch": 2.3745087029758563, |
| "grad_norm": 0.15951845049858093, |
| "learning_rate": 1.253436041376417e-06, |
| "loss": 0.4631, |
| "step": 528 |
| }, |
| { |
| "epoch": 2.3790005614823135, |
| "grad_norm": 0.1614406406879425, |
| "learning_rate": 1.2361219262509882e-06, |
| "loss": 0.4507, |
| "step": 529 |
| }, |
| { |
| "epoch": 2.3834924199887704, |
| "grad_norm": 0.162693589925766, |
| "learning_rate": 1.2189113445673528e-06, |
| "loss": 0.4686, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.387984278495227, |
| "grad_norm": 0.1591520756483078, |
| "learning_rate": 1.201804769739028e-06, |
| "loss": 0.4288, |
| "step": 531 |
| }, |
| { |
| "epoch": 2.3924761370016845, |
| "grad_norm": 0.1632116734981537, |
| "learning_rate": 1.1848026723186013e-06, |
| "loss": 0.4627, |
| "step": 532 |
| }, |
| { |
| "epoch": 2.3969679955081413, |
| "grad_norm": 0.15154661238193512, |
| "learning_rate": 1.1679055199847894e-06, |
| "loss": 0.4392, |
| "step": 533 |
| }, |
| { |
| "epoch": 2.4014598540145986, |
| "grad_norm": 0.15290887653827667, |
| "learning_rate": 1.1511137775295705e-06, |
| "loss": 0.4468, |
| "step": 534 |
| }, |
| { |
| "epoch": 2.4059517125210554, |
| "grad_norm": 0.15557996928691864, |
| "learning_rate": 1.1344279068454012e-06, |
| "loss": 0.4629, |
| "step": 535 |
| }, |
| { |
| "epoch": 2.4104435710275127, |
| "grad_norm": 0.15348927676677704, |
| "learning_rate": 1.1178483669125112e-06, |
| "loss": 0.4421, |
| "step": 536 |
| }, |
| { |
| "epoch": 2.4149354295339696, |
| "grad_norm": 0.14844442903995514, |
| "learning_rate": 1.101375613786278e-06, |
| "loss": 0.4501, |
| "step": 537 |
| }, |
| { |
| "epoch": 2.419427288040427, |
| "grad_norm": 0.15993905067443848, |
| "learning_rate": 1.0850101005846787e-06, |
| "loss": 0.4595, |
| "step": 538 |
| }, |
| { |
| "epoch": 2.4239191465468837, |
| "grad_norm": 0.14879821240901947, |
| "learning_rate": 1.0687522774758318e-06, |
| "loss": 0.4146, |
| "step": 539 |
| }, |
| { |
| "epoch": 2.428411005053341, |
| "grad_norm": 0.14690467715263367, |
| "learning_rate": 1.052602591665612e-06, |
| "loss": 0.4579, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.432902863559798, |
| "grad_norm": 0.1457294076681137, |
| "learning_rate": 1.0365614873853463e-06, |
| "loss": 0.4484, |
| "step": 541 |
| }, |
| { |
| "epoch": 2.437394722066255, |
| "grad_norm": 0.15229414403438568, |
| "learning_rate": 1.0206294058795974e-06, |
| "loss": 0.4357, |
| "step": 542 |
| }, |
| { |
| "epoch": 2.441886580572712, |
| "grad_norm": 0.15288710594177246, |
| "learning_rate": 1.0048067853940286e-06, |
| "loss": 0.4415, |
| "step": 543 |
| }, |
| { |
| "epoch": 2.446378439079169, |
| "grad_norm": 0.15801645815372467, |
| "learning_rate": 9.890940611633416e-07, |
| "loss": 0.4452, |
| "step": 544 |
| }, |
| { |
| "epoch": 2.450870297585626, |
| "grad_norm": 0.14208218455314636, |
| "learning_rate": 9.734916653993104e-07, |
| "loss": 0.4478, |
| "step": 545 |
| }, |
| { |
| "epoch": 2.4553621560920833, |
| "grad_norm": 0.14192849397659302, |
| "learning_rate": 9.580000272788915e-07, |
| "loss": 0.4202, |
| "step": 546 |
| }, |
| { |
| "epoch": 2.45985401459854, |
| "grad_norm": 0.14638514816761017, |
| "learning_rate": 9.426195729324161e-07, |
| "loss": 0.4462, |
| "step": 547 |
| }, |
| { |
| "epoch": 2.464345873104997, |
| "grad_norm": 0.1476951539516449, |
| "learning_rate": 9.27350725431872e-07, |
| "loss": 0.4697, |
| "step": 548 |
| }, |
| { |
| "epoch": 2.4688377316114543, |
| "grad_norm": 0.13385483622550964, |
| "learning_rate": 9.121939047792622e-07, |
| "loss": 0.4335, |
| "step": 549 |
| }, |
| { |
| "epoch": 2.473329590117911, |
| "grad_norm": 0.15053848922252655, |
| "learning_rate": 8.971495278950559e-07, |
| "loss": 0.4633, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.4778214486243684, |
| "grad_norm": 0.14375874400138855, |
| "learning_rate": 8.822180086067161e-07, |
| "loss": 0.4618, |
| "step": 551 |
| }, |
| { |
| "epoch": 2.482313307130825, |
| "grad_norm": 0.14123788475990295, |
| "learning_rate": 8.673997576373205e-07, |
| "loss": 0.441, |
| "step": 552 |
| }, |
| { |
| "epoch": 2.4868051656372825, |
| "grad_norm": 0.1446654200553894, |
| "learning_rate": 8.526951825942608e-07, |
| "loss": 0.4495, |
| "step": 553 |
| }, |
| { |
| "epoch": 2.4912970241437393, |
| "grad_norm": 0.14551837742328644, |
| "learning_rate": 8.381046879580307e-07, |
| "loss": 0.437, |
| "step": 554 |
| }, |
| { |
| "epoch": 2.4957888826501966, |
| "grad_norm": 0.13555008172988892, |
| "learning_rate": 8.23628675071102e-07, |
| "loss": 0.4535, |
| "step": 555 |
| }, |
| { |
| "epoch": 2.5002807411566534, |
| "grad_norm": 0.1343921571969986, |
| "learning_rate": 8.092675421268825e-07, |
| "loss": 0.4214, |
| "step": 556 |
| }, |
| { |
| "epoch": 2.5047725996631107, |
| "grad_norm": 0.14767403900623322, |
| "learning_rate": 7.950216841587638e-07, |
| "loss": 0.4737, |
| "step": 557 |
| }, |
| { |
| "epoch": 2.5092644581695676, |
| "grad_norm": 0.13317753374576569, |
| "learning_rate": 7.808914930292544e-07, |
| "loss": 0.4147, |
| "step": 558 |
| }, |
| { |
| "epoch": 2.513756316676025, |
| "grad_norm": 0.14309468865394592, |
| "learning_rate": 7.66877357419204e-07, |
| "loss": 0.4357, |
| "step": 559 |
| }, |
| { |
| "epoch": 2.5182481751824817, |
| "grad_norm": 0.1514693796634674, |
| "learning_rate": 7.529796628171071e-07, |
| "loss": 0.4626, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.522740033688939, |
| "grad_norm": 0.14453499019145966, |
| "learning_rate": 7.391987915085014e-07, |
| "loss": 0.4349, |
| "step": 561 |
| }, |
| { |
| "epoch": 2.527231892195396, |
| "grad_norm": 0.14623714983463287, |
| "learning_rate": 7.255351225654527e-07, |
| "loss": 0.4436, |
| "step": 562 |
| }, |
| { |
| "epoch": 2.531723750701853, |
| "grad_norm": 0.14449186623096466, |
| "learning_rate": 7.119890318361278e-07, |
| "loss": 0.4318, |
| "step": 563 |
| }, |
| { |
| "epoch": 2.53621560920831, |
| "grad_norm": 0.15518885850906372, |
| "learning_rate": 6.98560891934455e-07, |
| "loss": 0.4568, |
| "step": 564 |
| }, |
| { |
| "epoch": 2.5407074677147667, |
| "grad_norm": 0.15409418940544128, |
| "learning_rate": 6.852510722298761e-07, |
| "loss": 0.4654, |
| "step": 565 |
| }, |
| { |
| "epoch": 2.545199326221224, |
| "grad_norm": 0.1519947648048401, |
| "learning_rate": 6.72059938837184e-07, |
| "loss": 0.4567, |
| "step": 566 |
| }, |
| { |
| "epoch": 2.5496911847276813, |
| "grad_norm": 0.13835862278938293, |
| "learning_rate": 6.589878546064544e-07, |
| "loss": 0.4312, |
| "step": 567 |
| }, |
| { |
| "epoch": 2.554183043234138, |
| "grad_norm": 0.1384487897157669, |
| "learning_rate": 6.46035179113062e-07, |
| "loss": 0.4264, |
| "step": 568 |
| }, |
| { |
| "epoch": 2.558674901740595, |
| "grad_norm": 0.16356123983860016, |
| "learning_rate": 6.332022686477929e-07, |
| "loss": 0.4936, |
| "step": 569 |
| }, |
| { |
| "epoch": 2.5631667602470523, |
| "grad_norm": 0.14139850437641144, |
| "learning_rate": 6.204894762070407e-07, |
| "loss": 0.4379, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.5676586187535095, |
| "grad_norm": 0.14354942739009857, |
| "learning_rate": 6.078971514830989e-07, |
| "loss": 0.4396, |
| "step": 571 |
| }, |
| { |
| "epoch": 2.5721504772599664, |
| "grad_norm": 0.17011038959026337, |
| "learning_rate": 5.954256408545417e-07, |
| "loss": 0.4474, |
| "step": 572 |
| }, |
| { |
| "epoch": 2.576642335766423, |
| "grad_norm": 0.13311436772346497, |
| "learning_rate": 5.830752873766948e-07, |
| "loss": 0.4325, |
| "step": 573 |
| }, |
| { |
| "epoch": 2.5811341942728805, |
| "grad_norm": 0.14901982247829437, |
| "learning_rate": 5.708464307722006e-07, |
| "loss": 0.481, |
| "step": 574 |
| }, |
| { |
| "epoch": 2.5856260527793373, |
| "grad_norm": 0.14402733743190765, |
| "learning_rate": 5.587394074216712e-07, |
| "loss": 0.4279, |
| "step": 575 |
| }, |
| { |
| "epoch": 2.5901179112857946, |
| "grad_norm": 0.14105503261089325, |
| "learning_rate": 5.467545503544381e-07, |
| "loss": 0.4456, |
| "step": 576 |
| }, |
| { |
| "epoch": 2.5946097697922514, |
| "grad_norm": 0.15389291942119598, |
| "learning_rate": 5.348921892393905e-07, |
| "loss": 0.4538, |
| "step": 577 |
| }, |
| { |
| "epoch": 2.5991016282987087, |
| "grad_norm": 0.13729771971702576, |
| "learning_rate": 5.231526503759055e-07, |
| "loss": 0.4328, |
| "step": 578 |
| }, |
| { |
| "epoch": 2.6035934868051656, |
| "grad_norm": 0.1453271359205246, |
| "learning_rate": 5.115362566848747e-07, |
| "loss": 0.4522, |
| "step": 579 |
| }, |
| { |
| "epoch": 2.608085345311623, |
| "grad_norm": 0.1378541737794876, |
| "learning_rate": 5.000433276998218e-07, |
| "loss": 0.4461, |
| "step": 580 |
| }, |
| { |
| "epoch": 2.6125772038180797, |
| "grad_norm": 0.14585429430007935, |
| "learning_rate": 4.886741795581101e-07, |
| "loss": 0.4487, |
| "step": 581 |
| }, |
| { |
| "epoch": 2.6170690623245365, |
| "grad_norm": 0.13975073397159576, |
| "learning_rate": 4.774291249922508e-07, |
| "loss": 0.4603, |
| "step": 582 |
| }, |
| { |
| "epoch": 2.621560920830994, |
| "grad_norm": 0.130731463432312, |
| "learning_rate": 4.663084733212958e-07, |
| "loss": 0.4121, |
| "step": 583 |
| }, |
| { |
| "epoch": 2.626052779337451, |
| "grad_norm": 0.13513977825641632, |
| "learning_rate": 4.5531253044233393e-07, |
| "loss": 0.4589, |
| "step": 584 |
| }, |
| { |
| "epoch": 2.630544637843908, |
| "grad_norm": 0.13990791141986847, |
| "learning_rate": 4.4444159882207406e-07, |
| "loss": 0.4352, |
| "step": 585 |
| }, |
| { |
| "epoch": 2.6350364963503647, |
| "grad_norm": 0.1394883394241333, |
| "learning_rate": 4.336959774885241e-07, |
| "loss": 0.4662, |
| "step": 586 |
| }, |
| { |
| "epoch": 2.639528354856822, |
| "grad_norm": 0.14077690243721008, |
| "learning_rate": 4.230759620227681e-07, |
| "loss": 0.4468, |
| "step": 587 |
| }, |
| { |
| "epoch": 2.6440202133632793, |
| "grad_norm": 0.14740774035453796, |
| "learning_rate": 4.125818445508351e-07, |
| "loss": 0.4505, |
| "step": 588 |
| }, |
| { |
| "epoch": 2.648512071869736, |
| "grad_norm": 0.13872882723808289, |
| "learning_rate": 4.022139137356623e-07, |
| "loss": 0.432, |
| "step": 589 |
| }, |
| { |
| "epoch": 2.653003930376193, |
| "grad_norm": 0.1328064501285553, |
| "learning_rate": 3.9197245476915556e-07, |
| "loss": 0.4578, |
| "step": 590 |
| }, |
| { |
| "epoch": 2.6574957888826503, |
| "grad_norm": 0.1333392858505249, |
| "learning_rate": 3.818577493643444e-07, |
| "loss": 0.4516, |
| "step": 591 |
| }, |
| { |
| "epoch": 2.661987647389107, |
| "grad_norm": 0.13673578202724457, |
| "learning_rate": 3.718700757476323e-07, |
| "loss": 0.4443, |
| "step": 592 |
| }, |
| { |
| "epoch": 2.6664795058955644, |
| "grad_norm": 0.14025312662124634, |
| "learning_rate": 3.6200970865114705e-07, |
| "loss": 0.4367, |
| "step": 593 |
| }, |
| { |
| "epoch": 2.670971364402021, |
| "grad_norm": 0.13922296464443207, |
| "learning_rate": 3.5227691930517895e-07, |
| "loss": 0.4489, |
| "step": 594 |
| }, |
| { |
| "epoch": 2.6754632229084785, |
| "grad_norm": 0.1483432501554489, |
| "learning_rate": 3.426719754307206e-07, |
| "loss": 0.4535, |
| "step": 595 |
| }, |
| { |
| "epoch": 2.6799550814149353, |
| "grad_norm": 0.14136959612369537, |
| "learning_rate": 3.331951412321066e-07, |
| "loss": 0.4441, |
| "step": 596 |
| }, |
| { |
| "epoch": 2.6844469399213926, |
| "grad_norm": 0.14010727405548096, |
| "learning_rate": 3.23846677389742e-07, |
| "loss": 0.4508, |
| "step": 597 |
| }, |
| { |
| "epoch": 2.6889387984278494, |
| "grad_norm": 0.13651929795742035, |
| "learning_rate": 3.14626841052933e-07, |
| "loss": 0.4367, |
| "step": 598 |
| }, |
| { |
| "epoch": 2.6934306569343067, |
| "grad_norm": 0.1339380443096161, |
| "learning_rate": 3.0553588583281446e-07, |
| "loss": 0.4498, |
| "step": 599 |
| }, |
| { |
| "epoch": 2.6979225154407636, |
| "grad_norm": 0.14083103835582733, |
| "learning_rate": 2.965740617953733e-07, |
| "loss": 0.4276, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.702414373947221, |
| "grad_norm": 0.13285264372825623, |
| "learning_rate": 2.877416154545681e-07, |
| "loss": 0.4547, |
| "step": 601 |
| }, |
| { |
| "epoch": 2.7069062324536777, |
| "grad_norm": 0.13597218692302704, |
| "learning_rate": 2.7903878976555165e-07, |
| "loss": 0.4384, |
| "step": 602 |
| }, |
| { |
| "epoch": 2.7113980909601345, |
| "grad_norm": 0.14728978276252747, |
| "learning_rate": 2.704658241179847e-07, |
| "loss": 0.4478, |
| "step": 603 |
| }, |
| { |
| "epoch": 2.715889949466592, |
| "grad_norm": 0.1403026282787323, |
| "learning_rate": 2.620229543294528e-07, |
| "loss": 0.4463, |
| "step": 604 |
| }, |
| { |
| "epoch": 2.720381807973049, |
| "grad_norm": 0.14910593628883362, |
| "learning_rate": 2.5371041263897945e-07, |
| "loss": 0.4636, |
| "step": 605 |
| }, |
| { |
| "epoch": 2.724873666479506, |
| "grad_norm": 0.13558971881866455, |
| "learning_rate": 2.4552842770063756e-07, |
| "loss": 0.4223, |
| "step": 606 |
| }, |
| { |
| "epoch": 2.7293655249859627, |
| "grad_norm": 0.13463318347930908, |
| "learning_rate": 2.3747722457725996e-07, |
| "loss": 0.4413, |
| "step": 607 |
| }, |
| { |
| "epoch": 2.73385738349242, |
| "grad_norm": 0.12877213954925537, |
| "learning_rate": 2.2955702473424824e-07, |
| "loss": 0.4399, |
| "step": 608 |
| }, |
| { |
| "epoch": 2.7383492419988773, |
| "grad_norm": 0.13977070152759552, |
| "learning_rate": 2.217680460334809e-07, |
| "loss": 0.4518, |
| "step": 609 |
| }, |
| { |
| "epoch": 2.742841100505334, |
| "grad_norm": 0.13352471590042114, |
| "learning_rate": 2.141105027273227e-07, |
| "loss": 0.4456, |
| "step": 610 |
| }, |
| { |
| "epoch": 2.747332959011791, |
| "grad_norm": 0.14282134175300598, |
| "learning_rate": 2.0658460545272653e-07, |
| "loss": 0.4272, |
| "step": 611 |
| }, |
| { |
| "epoch": 2.7518248175182483, |
| "grad_norm": 0.13226774334907532, |
| "learning_rate": 1.9919056122544467e-07, |
| "loss": 0.446, |
| "step": 612 |
| }, |
| { |
| "epoch": 2.756316676024705, |
| "grad_norm": 0.13558903336524963, |
| "learning_rate": 1.9192857343433069e-07, |
| "loss": 0.441, |
| "step": 613 |
| }, |
| { |
| "epoch": 2.7608085345311624, |
| "grad_norm": 0.13000066578388214, |
| "learning_rate": 1.847988418357466e-07, |
| "loss": 0.4406, |
| "step": 614 |
| }, |
| { |
| "epoch": 2.765300393037619, |
| "grad_norm": 0.13514713943004608, |
| "learning_rate": 1.778015625480678e-07, |
| "loss": 0.4497, |
| "step": 615 |
| }, |
| { |
| "epoch": 2.7697922515440765, |
| "grad_norm": 0.1365610659122467, |
| "learning_rate": 1.7093692804628637e-07, |
| "loss": 0.4279, |
| "step": 616 |
| }, |
| { |
| "epoch": 2.7742841100505333, |
| "grad_norm": 0.1374463140964508, |
| "learning_rate": 1.642051271567213e-07, |
| "loss": 0.4638, |
| "step": 617 |
| }, |
| { |
| "epoch": 2.7787759685569906, |
| "grad_norm": 0.12901850044727325, |
| "learning_rate": 1.5760634505182004e-07, |
| "loss": 0.4493, |
| "step": 618 |
| }, |
| { |
| "epoch": 2.7832678270634474, |
| "grad_norm": 0.13684892654418945, |
| "learning_rate": 1.5114076324506567e-07, |
| "loss": 0.4433, |
| "step": 619 |
| }, |
| { |
| "epoch": 2.7877596855699043, |
| "grad_norm": 0.13286025822162628, |
| "learning_rate": 1.4480855958598716e-07, |
| "loss": 0.4239, |
| "step": 620 |
| }, |
| { |
| "epoch": 2.7922515440763616, |
| "grad_norm": 0.1406407654285431, |
| "learning_rate": 1.3860990825526334e-07, |
| "loss": 0.4582, |
| "step": 621 |
| }, |
| { |
| "epoch": 2.796743402582819, |
| "grad_norm": 0.12750087678432465, |
| "learning_rate": 1.3254497975993263e-07, |
| "loss": 0.4476, |
| "step": 622 |
| }, |
| { |
| "epoch": 2.8012352610892757, |
| "grad_norm": 0.14279097318649292, |
| "learning_rate": 1.266139409287054e-07, |
| "loss": 0.4591, |
| "step": 623 |
| }, |
| { |
| "epoch": 2.8057271195957325, |
| "grad_norm": 0.13092957437038422, |
| "learning_rate": 1.2081695490737177e-07, |
| "loss": 0.4373, |
| "step": 624 |
| }, |
| { |
| "epoch": 2.81021897810219, |
| "grad_norm": 0.14031663537025452, |
| "learning_rate": 1.1515418115431554e-07, |
| "loss": 0.4418, |
| "step": 625 |
| }, |
| { |
| "epoch": 2.814710836608647, |
| "grad_norm": 0.1413896381855011, |
| "learning_rate": 1.0962577543612796e-07, |
| "loss": 0.447, |
| "step": 626 |
| }, |
| { |
| "epoch": 2.819202695115104, |
| "grad_norm": 0.13287171721458435, |
| "learning_rate": 1.04231889823323e-07, |
| "loss": 0.4297, |
| "step": 627 |
| }, |
| { |
| "epoch": 2.8236945536215607, |
| "grad_norm": 0.13622640073299408, |
| "learning_rate": 9.897267268615285e-08, |
| "loss": 0.46, |
| "step": 628 |
| }, |
| { |
| "epoch": 2.828186412128018, |
| "grad_norm": 0.13213184475898743, |
| "learning_rate": 9.384826869052899e-08, |
| "loss": 0.4498, |
| "step": 629 |
| }, |
| { |
| "epoch": 2.832678270634475, |
| "grad_norm": 0.13575822114944458, |
| "learning_rate": 8.885881879404201e-08, |
| "loss": 0.4398, |
| "step": 630 |
| }, |
| { |
| "epoch": 2.837170129140932, |
| "grad_norm": 0.13295438885688782, |
| "learning_rate": 8.400446024208309e-08, |
| "loss": 0.4562, |
| "step": 631 |
| }, |
| { |
| "epoch": 2.841661987647389, |
| "grad_norm": 0.13092169165611267, |
| "learning_rate": 7.928532656407029e-08, |
| "loss": 0.4405, |
| "step": 632 |
| }, |
| { |
| "epoch": 2.8461538461538463, |
| "grad_norm": 0.1367148458957672, |
| "learning_rate": 7.470154756977544e-08, |
| "loss": 0.4483, |
| "step": 633 |
| }, |
| { |
| "epoch": 2.850645704660303, |
| "grad_norm": 0.13294926285743713, |
| "learning_rate": 7.02532493457514e-08, |
| "loss": 0.4552, |
| "step": 634 |
| }, |
| { |
| "epoch": 2.8551375631667604, |
| "grad_norm": 0.13394087553024292, |
| "learning_rate": 6.594055425186763e-08, |
| "loss": 0.4607, |
| "step": 635 |
| }, |
| { |
| "epoch": 2.859629421673217, |
| "grad_norm": 0.12872961163520813, |
| "learning_rate": 6.176358091794011e-08, |
| "loss": 0.4353, |
| "step": 636 |
| }, |
| { |
| "epoch": 2.8641212801796745, |
| "grad_norm": 0.13416117429733276, |
| "learning_rate": 5.772244424047169e-08, |
| "loss": 0.437, |
| "step": 637 |
| }, |
| { |
| "epoch": 2.8686131386861313, |
| "grad_norm": 0.13002049922943115, |
| "learning_rate": 5.3817255379488565e-08, |
| "loss": 0.4512, |
| "step": 638 |
| }, |
| { |
| "epoch": 2.8731049971925886, |
| "grad_norm": 0.14305950701236725, |
| "learning_rate": 5.004812175548657e-08, |
| "loss": 0.4604, |
| "step": 639 |
| }, |
| { |
| "epoch": 2.8775968556990454, |
| "grad_norm": 0.1293281465768814, |
| "learning_rate": 4.641514704647132e-08, |
| "loss": 0.4368, |
| "step": 640 |
| }, |
| { |
| "epoch": 2.8820887142055023, |
| "grad_norm": 0.13169856369495392, |
| "learning_rate": 4.2918431185110523e-08, |
| "loss": 0.4241, |
| "step": 641 |
| }, |
| { |
| "epoch": 2.8865805727119596, |
| "grad_norm": 0.13361629843711853, |
| "learning_rate": 3.9558070355983356e-08, |
| "loss": 0.4689, |
| "step": 642 |
| }, |
| { |
| "epoch": 2.891072431218417, |
| "grad_norm": 0.13451144099235535, |
| "learning_rate": 3.633415699293541e-08, |
| "loss": 0.4431, |
| "step": 643 |
| }, |
| { |
| "epoch": 2.8955642897248737, |
| "grad_norm": 0.1316562294960022, |
| "learning_rate": 3.324677977653401e-08, |
| "loss": 0.4403, |
| "step": 644 |
| }, |
| { |
| "epoch": 2.9000561482313305, |
| "grad_norm": 0.12996108829975128, |
| "learning_rate": 3.0296023631631866e-08, |
| "loss": 0.4472, |
| "step": 645 |
| }, |
| { |
| "epoch": 2.904548006737788, |
| "grad_norm": 0.13493524491786957, |
| "learning_rate": 2.7481969725028923e-08, |
| "loss": 0.4486, |
| "step": 646 |
| }, |
| { |
| "epoch": 2.909039865244245, |
| "grad_norm": 0.12959401309490204, |
| "learning_rate": 2.4804695463240825e-08, |
| "loss": 0.435, |
| "step": 647 |
| }, |
| { |
| "epoch": 2.913531723750702, |
| "grad_norm": 0.13398133218288422, |
| "learning_rate": 2.226427449036894e-08, |
| "loss": 0.4465, |
| "step": 648 |
| }, |
| { |
| "epoch": 2.9180235822571587, |
| "grad_norm": 0.12995387613773346, |
| "learning_rate": 1.9860776686075332e-08, |
| "loss": 0.4491, |
| "step": 649 |
| }, |
| { |
| "epoch": 2.922515440763616, |
| "grad_norm": 0.1356271654367447, |
| "learning_rate": 1.7594268163659277e-08, |
| "loss": 0.4449, |
| "step": 650 |
| }, |
| { |
| "epoch": 2.927007299270073, |
| "grad_norm": 0.12999212741851807, |
| "learning_rate": 1.546481126824151e-08, |
| "loss": 0.4446, |
| "step": 651 |
| }, |
| { |
| "epoch": 2.93149915777653, |
| "grad_norm": 0.1264127790927887, |
| "learning_rate": 1.347246457504503e-08, |
| "loss": 0.4397, |
| "step": 652 |
| }, |
| { |
| "epoch": 2.935991016282987, |
| "grad_norm": 0.13195131719112396, |
| "learning_rate": 1.1617282887787517e-08, |
| "loss": 0.4434, |
| "step": 653 |
| }, |
| { |
| "epoch": 2.9404828747894443, |
| "grad_norm": 0.13151754438877106, |
| "learning_rate": 9.899317237172524e-09, |
| "loss": 0.432, |
| "step": 654 |
| }, |
| { |
| "epoch": 2.944974733295901, |
| "grad_norm": 0.1297924965620041, |
| "learning_rate": 8.318614879485044e-09, |
| "loss": 0.4271, |
| "step": 655 |
| }, |
| { |
| "epoch": 2.9494665918023584, |
| "grad_norm": 0.1355123370885849, |
| "learning_rate": 6.8752192952931115e-09, |
| "loss": 0.4666, |
| "step": 656 |
| }, |
| { |
| "epoch": 2.953958450308815, |
| "grad_norm": 0.13056360185146332, |
| "learning_rate": 5.569170188250983e-09, |
| "loss": 0.4374, |
| "step": 657 |
| }, |
| { |
| "epoch": 2.958450308815272, |
| "grad_norm": 0.15338468551635742, |
| "learning_rate": 4.4005034840061135e-09, |
| "loss": 0.4453, |
| "step": 658 |
| }, |
| { |
| "epoch": 2.9629421673217293, |
| "grad_norm": 0.13862362504005432, |
| "learning_rate": 3.3692513292132855e-09, |
| "loss": 0.4584, |
| "step": 659 |
| }, |
| { |
| "epoch": 2.9674340258281866, |
| "grad_norm": 0.13754971325397491, |
| "learning_rate": 2.4754420906475396e-09, |
| "loss": 0.4386, |
| "step": 660 |
| }, |
| { |
| "epoch": 2.9719258843346434, |
| "grad_norm": 0.13375526666641235, |
| "learning_rate": 1.7191003544259067e-09, |
| "loss": 0.4451, |
| "step": 661 |
| }, |
| { |
| "epoch": 2.9764177428411003, |
| "grad_norm": 0.13616608083248138, |
| "learning_rate": 1.100246925331283e-09, |
| "loss": 0.4446, |
| "step": 662 |
| }, |
| { |
| "epoch": 2.9809096013475576, |
| "grad_norm": 0.1345696747303009, |
| "learning_rate": 6.188988262373353e-10, |
| "loss": 0.4473, |
| "step": 663 |
| }, |
| { |
| "epoch": 2.985401459854015, |
| "grad_norm": 0.13510431349277496, |
| "learning_rate": 2.750692976444258e-10, |
| "loss": 0.4373, |
| "step": 664 |
| }, |
| { |
| "epoch": 2.9898933183604717, |
| "grad_norm": 0.13065224885940552, |
| "learning_rate": 6.876779731213035e-11, |
| "loss": 0.4428, |
| "step": 665 |
| }, |
| { |
| "epoch": 2.9943851768669285, |
| "grad_norm": 0.1353355348110199, |
| "learning_rate": 0.0, |
| "loss": 0.4488, |
| "step": 666 |
| }, |
| { |
| "epoch": 2.9943851768669285, |
| "step": 666, |
| "total_flos": 4607052148899840.0, |
| "train_loss": 0.4853614707429846, |
| "train_runtime": 116828.6857, |
| "train_samples_per_second": 2.926, |
| "train_steps_per_second": 0.006 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 666, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4607052148899840.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|