{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 338, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002962962962962963, "grad_norm": 82.5, "learning_rate": 0.0, "loss": 5.4851, "step": 1 }, { "epoch": 0.005925925925925926, "grad_norm": 81.5, "learning_rate": 1.0000000000000002e-06, "loss": 5.5219, "step": 2 }, { "epoch": 0.008888888888888889, "grad_norm": 83.0, "learning_rate": 2.0000000000000003e-06, "loss": 5.7155, "step": 3 }, { "epoch": 0.011851851851851851, "grad_norm": 74.0, "learning_rate": 3e-06, "loss": 5.5065, "step": 4 }, { "epoch": 0.014814814814814815, "grad_norm": 60.5, "learning_rate": 4.000000000000001e-06, "loss": 4.8696, "step": 5 }, { "epoch": 0.017777777777777778, "grad_norm": 48.25, "learning_rate": 5e-06, "loss": 4.3835, "step": 6 }, { "epoch": 0.02074074074074074, "grad_norm": 40.5, "learning_rate": 6e-06, "loss": 4.1231, "step": 7 }, { "epoch": 0.023703703703703703, "grad_norm": 50.5, "learning_rate": 7e-06, "loss": 3.6886, "step": 8 }, { "epoch": 0.02666666666666667, "grad_norm": 34.5, "learning_rate": 8.000000000000001e-06, "loss": 3.3401, "step": 9 }, { "epoch": 0.02962962962962963, "grad_norm": 27.125, "learning_rate": 9e-06, "loss": 3.043, "step": 10 }, { "epoch": 0.03259259259259259, "grad_norm": 21.5, "learning_rate": 1e-05, "loss": 2.7408, "step": 11 }, { "epoch": 0.035555555555555556, "grad_norm": 19.875, "learning_rate": 1.1000000000000001e-05, "loss": 2.5235, "step": 12 }, { "epoch": 0.03851851851851852, "grad_norm": 19.0, "learning_rate": 1.2e-05, "loss": 2.4532, "step": 13 }, { "epoch": 0.04148148148148148, "grad_norm": 19.625, "learning_rate": 1.3000000000000001e-05, "loss": 2.3238, "step": 14 }, { "epoch": 0.044444444444444446, "grad_norm": 20.125, "learning_rate": 1.4e-05, "loss": 2.2851, "step": 15 }, { "epoch": 0.047407407407407405, "grad_norm": 17.5, "learning_rate": 1.5000000000000002e-05, "loss": 2.2676, "step": 16 }, { "epoch": 0.05037037037037037, "grad_norm": 16.0, "learning_rate": 1.6000000000000003e-05, "loss": 2.1502, "step": 17 }, { "epoch": 0.05333333333333334, "grad_norm": 14.4375, "learning_rate": 1.7e-05, "loss": 2.1293, "step": 18 }, { "epoch": 0.056296296296296296, "grad_norm": 15.5625, "learning_rate": 1.8e-05, "loss": 2.0649, "step": 19 }, { "epoch": 0.05925925925925926, "grad_norm": 19.0, "learning_rate": 1.9e-05, "loss": 1.961, "step": 20 }, { "epoch": 0.06222222222222222, "grad_norm": 14.375, "learning_rate": 2e-05, "loss": 1.9368, "step": 21 }, { "epoch": 0.06518518518518518, "grad_norm": 15.75, "learning_rate": 1.999951200881446e-05, "loss": 1.9511, "step": 22 }, { "epoch": 0.06814814814814815, "grad_norm": 13.5625, "learning_rate": 1.999804808288491e-05, "loss": 1.8874, "step": 23 }, { "epoch": 0.07111111111111111, "grad_norm": 13.4375, "learning_rate": 1.9995608365087945e-05, "loss": 1.7678, "step": 24 }, { "epoch": 0.07407407407407407, "grad_norm": 14.25, "learning_rate": 1.999219309353572e-05, "loss": 1.8061, "step": 25 }, { "epoch": 0.07703703703703704, "grad_norm": 13.9375, "learning_rate": 1.9987802601552717e-05, "loss": 1.778, "step": 26 }, { "epoch": 0.08, "grad_norm": 11.3125, "learning_rate": 1.9982437317643218e-05, "loss": 1.5772, "step": 27 }, { "epoch": 0.08296296296296296, "grad_norm": 13.3125, "learning_rate": 1.9976097765449463e-05, "loss": 1.7557, "step": 28 }, { "epoch": 0.08592592592592592, "grad_norm": 12.9375, "learning_rate": 1.9968784563700586e-05, "loss": 1.6816, "step": 29 }, { "epoch": 0.08888888888888889, "grad_norm": 12.8125, "learning_rate": 1.996049842615217e-05, "loss": 1.742, "step": 30 }, { "epoch": 0.09185185185185185, "grad_norm": 12.125, "learning_rate": 1.9951240161516643e-05, "loss": 1.7055, "step": 31 }, { "epoch": 0.09481481481481481, "grad_norm": 11.5625, "learning_rate": 1.9941010673384307e-05, "loss": 1.6907, "step": 32 }, { "epoch": 0.09777777777777778, "grad_norm": 11.4375, "learning_rate": 1.992981096013517e-05, "loss": 1.7213, "step": 33 }, { "epoch": 0.10074074074074074, "grad_norm": 12.125, "learning_rate": 1.9917642114841505e-05, "loss": 1.8787, "step": 34 }, { "epoch": 0.1037037037037037, "grad_norm": 11.125, "learning_rate": 1.990450532516116e-05, "loss": 1.6781, "step": 35 }, { "epoch": 0.10666666666666667, "grad_norm": 12.8125, "learning_rate": 1.9890401873221642e-05, "loss": 1.5917, "step": 36 }, { "epoch": 0.10962962962962963, "grad_norm": 12.0625, "learning_rate": 1.9875333135495e-05, "loss": 1.7399, "step": 37 }, { "epoch": 0.11259259259259259, "grad_norm": 11.6875, "learning_rate": 1.985930058266348e-05, "loss": 1.8376, "step": 38 }, { "epoch": 0.11555555555555555, "grad_norm": 10.9375, "learning_rate": 1.984230577947597e-05, "loss": 1.7302, "step": 39 }, { "epoch": 0.11851851851851852, "grad_norm": 10.125, "learning_rate": 1.9824350384595295e-05, "loss": 1.6194, "step": 40 }, { "epoch": 0.12148148148148148, "grad_norm": 10.375, "learning_rate": 1.9805436150436352e-05, "loss": 1.6648, "step": 41 }, { "epoch": 0.12444444444444444, "grad_norm": 12.3125, "learning_rate": 1.9785564922995042e-05, "loss": 1.7417, "step": 42 }, { "epoch": 0.1274074074074074, "grad_norm": 11.625, "learning_rate": 1.9764738641668137e-05, "loss": 1.6403, "step": 43 }, { "epoch": 0.13037037037037036, "grad_norm": 10.625, "learning_rate": 1.9742959339063977e-05, "loss": 1.6204, "step": 44 }, { "epoch": 0.13333333333333333, "grad_norm": 11.75, "learning_rate": 1.972022914080411e-05, "loss": 1.7575, "step": 45 }, { "epoch": 0.1362962962962963, "grad_norm": 12.625, "learning_rate": 1.9696550265315805e-05, "loss": 1.594, "step": 46 }, { "epoch": 0.13925925925925925, "grad_norm": 11.3125, "learning_rate": 1.9671925023615572e-05, "loss": 1.6259, "step": 47 }, { "epoch": 0.14222222222222222, "grad_norm": 11.9375, "learning_rate": 1.964635581908359e-05, "loss": 1.7574, "step": 48 }, { "epoch": 0.1451851851851852, "grad_norm": 10.75, "learning_rate": 1.961984514722914e-05, "loss": 1.7021, "step": 49 }, { "epoch": 0.14814814814814814, "grad_norm": 10.625, "learning_rate": 1.9592395595447064e-05, "loss": 1.5892, "step": 50 }, { "epoch": 0.1511111111111111, "grad_norm": 10.625, "learning_rate": 1.9564009842765225e-05, "loss": 1.6846, "step": 51 }, { "epoch": 0.15407407407407409, "grad_norm": 10.375, "learning_rate": 1.9534690659583045e-05, "loss": 1.6761, "step": 52 }, { "epoch": 0.15703703703703703, "grad_norm": 10.625, "learning_rate": 1.9504440907401113e-05, "loss": 1.4986, "step": 53 }, { "epoch": 0.16, "grad_norm": 11.4375, "learning_rate": 1.9473263538541916e-05, "loss": 1.6183, "step": 54 }, { "epoch": 0.16296296296296298, "grad_norm": 12.3125, "learning_rate": 1.944116159586169e-05, "loss": 1.7405, "step": 55 }, { "epoch": 0.16592592592592592, "grad_norm": 11.75, "learning_rate": 1.9408138212453456e-05, "loss": 1.608, "step": 56 }, { "epoch": 0.1688888888888889, "grad_norm": 10.6875, "learning_rate": 1.9374196611341212e-05, "loss": 1.6267, "step": 57 }, { "epoch": 0.17185185185185184, "grad_norm": 12.3125, "learning_rate": 1.933934010516539e-05, "loss": 1.7335, "step": 58 }, { "epoch": 0.1748148148148148, "grad_norm": 11.125, "learning_rate": 1.9303572095859545e-05, "loss": 1.7112, "step": 59 }, { "epoch": 0.17777777777777778, "grad_norm": 11.75, "learning_rate": 1.9266896074318335e-05, "loss": 1.8106, "step": 60 }, { "epoch": 0.18074074074074073, "grad_norm": 10.875, "learning_rate": 1.9229315620056805e-05, "loss": 1.5596, "step": 61 }, { "epoch": 0.1837037037037037, "grad_norm": 14.0625, "learning_rate": 1.9190834400861035e-05, "loss": 1.5379, "step": 62 }, { "epoch": 0.18666666666666668, "grad_norm": 10.875, "learning_rate": 1.9151456172430186e-05, "loss": 1.5921, "step": 63 }, { "epoch": 0.18962962962962962, "grad_norm": 10.75, "learning_rate": 1.9111184778009934e-05, "loss": 1.6252, "step": 64 }, { "epoch": 0.1925925925925926, "grad_norm": 11.8125, "learning_rate": 1.9070024148017375e-05, "loss": 1.7589, "step": 65 }, { "epoch": 0.19555555555555557, "grad_norm": 10.4375, "learning_rate": 1.9027978299657436e-05, "loss": 1.4877, "step": 66 }, { "epoch": 0.1985185185185185, "grad_norm": 10.0625, "learning_rate": 1.89850513365308e-05, "loss": 1.6578, "step": 67 }, { "epoch": 0.20148148148148148, "grad_norm": 11.5, "learning_rate": 1.8941247448233386e-05, "loss": 1.5563, "step": 68 }, { "epoch": 0.20444444444444446, "grad_norm": 12.25, "learning_rate": 1.8896570909947477e-05, "loss": 1.6964, "step": 69 }, { "epoch": 0.2074074074074074, "grad_norm": 10.5625, "learning_rate": 1.885102608202444e-05, "loss": 1.5913, "step": 70 }, { "epoch": 0.21037037037037037, "grad_norm": 10.625, "learning_rate": 1.88046174095592e-05, "loss": 1.5556, "step": 71 }, { "epoch": 0.21333333333333335, "grad_norm": 10.0625, "learning_rate": 1.875734942195637e-05, "loss": 1.5116, "step": 72 }, { "epoch": 0.2162962962962963, "grad_norm": 9.6875, "learning_rate": 1.8709226732488216e-05, "loss": 1.5118, "step": 73 }, { "epoch": 0.21925925925925926, "grad_norm": 10.4375, "learning_rate": 1.866025403784439e-05, "loss": 1.6544, "step": 74 }, { "epoch": 0.2222222222222222, "grad_norm": 10.875, "learning_rate": 1.8610436117673557e-05, "loss": 1.5481, "step": 75 }, { "epoch": 0.22518518518518518, "grad_norm": 10.75, "learning_rate": 1.8559777834116906e-05, "loss": 1.5817, "step": 76 }, { "epoch": 0.22814814814814816, "grad_norm": 10.5, "learning_rate": 1.8508284131333604e-05, "loss": 1.5016, "step": 77 }, { "epoch": 0.2311111111111111, "grad_norm": 10.8125, "learning_rate": 1.845596003501826e-05, "loss": 1.5924, "step": 78 }, { "epoch": 0.23407407407407407, "grad_norm": 10.25, "learning_rate": 1.8402810651910444e-05, "loss": 1.63, "step": 79 }, { "epoch": 0.23703703703703705, "grad_norm": 10.5625, "learning_rate": 1.8348841169296247e-05, "loss": 1.5974, "step": 80 }, { "epoch": 0.24, "grad_norm": 9.875, "learning_rate": 1.829405685450202e-05, "loss": 1.5722, "step": 81 }, { "epoch": 0.24296296296296296, "grad_norm": 10.5, "learning_rate": 1.823846305438032e-05, "loss": 1.4561, "step": 82 }, { "epoch": 0.24592592592592594, "grad_norm": 9.5625, "learning_rate": 1.8182065194788024e-05, "loss": 1.4567, "step": 83 }, { "epoch": 0.24888888888888888, "grad_norm": 9.875, "learning_rate": 1.8124868780056814e-05, "loss": 1.613, "step": 84 }, { "epoch": 0.2518518518518518, "grad_norm": 9.75, "learning_rate": 1.8066879392455932e-05, "loss": 1.609, "step": 85 }, { "epoch": 0.2548148148148148, "grad_norm": 10.1875, "learning_rate": 1.800810269164738e-05, "loss": 1.6807, "step": 86 }, { "epoch": 0.2577777777777778, "grad_norm": 10.3125, "learning_rate": 1.7948544414133534e-05, "loss": 1.6024, "step": 87 }, { "epoch": 0.2607407407407407, "grad_norm": 9.9375, "learning_rate": 1.7888210372697292e-05, "loss": 1.4733, "step": 88 }, { "epoch": 0.2637037037037037, "grad_norm": 10.4375, "learning_rate": 1.782710645583473e-05, "loss": 1.5994, "step": 89 }, { "epoch": 0.26666666666666666, "grad_norm": 13.625, "learning_rate": 1.7765238627180424e-05, "loss": 1.3751, "step": 90 }, { "epoch": 0.2696296296296296, "grad_norm": 10.875, "learning_rate": 1.7702612924925377e-05, "loss": 1.436, "step": 91 }, { "epoch": 0.2725925925925926, "grad_norm": 9.625, "learning_rate": 1.7639235461227727e-05, "loss": 1.5887, "step": 92 }, { "epoch": 0.27555555555555555, "grad_norm": 9.6875, "learning_rate": 1.7575112421616203e-05, "loss": 1.5977, "step": 93 }, { "epoch": 0.2785185185185185, "grad_norm": 10.25, "learning_rate": 1.751025006438643e-05, "loss": 1.518, "step": 94 }, { "epoch": 0.2814814814814815, "grad_norm": 10.8125, "learning_rate": 1.7444654719990128e-05, "loss": 1.6543, "step": 95 }, { "epoch": 0.28444444444444444, "grad_norm": 9.9375, "learning_rate": 1.7378332790417275e-05, "loss": 1.5368, "step": 96 }, { "epoch": 0.2874074074074074, "grad_norm": 10.0, "learning_rate": 1.7311290748571273e-05, "loss": 1.5349, "step": 97 }, { "epoch": 0.2903703703703704, "grad_norm": 9.875, "learning_rate": 1.7243535137637227e-05, "loss": 1.6126, "step": 98 }, { "epoch": 0.29333333333333333, "grad_norm": 9.8125, "learning_rate": 1.717507257044331e-05, "loss": 1.5037, "step": 99 }, { "epoch": 0.2962962962962963, "grad_norm": 9.3125, "learning_rate": 1.7105909728815398e-05, "loss": 1.5038, "step": 100 }, { "epoch": 0.2992592592592593, "grad_norm": 9.0, "learning_rate": 1.7036053362924896e-05, "loss": 1.3771, "step": 101 }, { "epoch": 0.3022222222222222, "grad_norm": 9.375, "learning_rate": 1.6965510290629973e-05, "loss": 1.5432, "step": 102 }, { "epoch": 0.30518518518518517, "grad_norm": 9.375, "learning_rate": 1.689428739681012e-05, "loss": 1.4479, "step": 103 }, { "epoch": 0.30814814814814817, "grad_norm": 11.1875, "learning_rate": 1.682239163269422e-05, "loss": 1.5956, "step": 104 }, { "epoch": 0.3111111111111111, "grad_norm": 10.5, "learning_rate": 1.6749830015182106e-05, "loss": 1.4063, "step": 105 }, { "epoch": 0.31407407407407406, "grad_norm": 10.0625, "learning_rate": 1.667660962615973e-05, "loss": 1.6329, "step": 106 }, { "epoch": 0.31703703703703706, "grad_norm": 11.3125, "learning_rate": 1.6602737611807975e-05, "loss": 1.5915, "step": 107 }, { "epoch": 0.32, "grad_norm": 9.375, "learning_rate": 1.6528221181905217e-05, "loss": 1.3679, "step": 108 }, { "epoch": 0.32296296296296295, "grad_norm": 9.1875, "learning_rate": 1.6453067609123656e-05, "loss": 1.4718, "step": 109 }, { "epoch": 0.32592592592592595, "grad_norm": 9.6875, "learning_rate": 1.6377284228319496e-05, "loss": 1.4929, "step": 110 }, { "epoch": 0.3288888888888889, "grad_norm": 9.9375, "learning_rate": 1.6300878435817115e-05, "loss": 1.4887, "step": 111 }, { "epoch": 0.33185185185185184, "grad_norm": 10.1875, "learning_rate": 1.622385768868716e-05, "loss": 1.5634, "step": 112 }, { "epoch": 0.3348148148148148, "grad_norm": 9.75, "learning_rate": 1.6146229504018777e-05, "loss": 1.4839, "step": 113 }, { "epoch": 0.3377777777777778, "grad_norm": 10.9375, "learning_rate": 1.6068001458185934e-05, "loss": 1.4955, "step": 114 }, { "epoch": 0.34074074074074073, "grad_norm": 9.5625, "learning_rate": 1.5989181186108003e-05, "loss": 1.4464, "step": 115 }, { "epoch": 0.3437037037037037, "grad_norm": 9.8125, "learning_rate": 1.5909776380504583e-05, "loss": 1.4568, "step": 116 }, { "epoch": 0.3466666666666667, "grad_norm": 10.4375, "learning_rate": 1.5829794791144723e-05, "loss": 1.5491, "step": 117 }, { "epoch": 0.3496296296296296, "grad_norm": 10.9375, "learning_rate": 1.5749244224090537e-05, "loss": 1.4774, "step": 118 }, { "epoch": 0.35259259259259257, "grad_norm": 10.875, "learning_rate": 1.566813254093538e-05, "loss": 1.5642, "step": 119 }, { "epoch": 0.35555555555555557, "grad_norm": 10.4375, "learning_rate": 1.5586467658036526e-05, "loss": 1.5229, "step": 120 }, { "epoch": 0.3585185185185185, "grad_norm": 10.1875, "learning_rate": 1.5504257545742585e-05, "loss": 1.5067, "step": 121 }, { "epoch": 0.36148148148148146, "grad_norm": 9.6875, "learning_rate": 1.5421510227615587e-05, "loss": 1.4074, "step": 122 }, { "epoch": 0.36444444444444446, "grad_norm": 11.3125, "learning_rate": 1.533823377964791e-05, "loss": 1.4726, "step": 123 }, { "epoch": 0.3674074074074074, "grad_norm": 12.75, "learning_rate": 1.5254436329474062e-05, "loss": 1.351, "step": 124 }, { "epoch": 0.37037037037037035, "grad_norm": 10.6875, "learning_rate": 1.517012605557746e-05, "loss": 1.5042, "step": 125 }, { "epoch": 0.37333333333333335, "grad_norm": 9.8125, "learning_rate": 1.5085311186492206e-05, "loss": 1.5669, "step": 126 }, { "epoch": 0.3762962962962963, "grad_norm": 9.75, "learning_rate": 1.5000000000000002e-05, "loss": 1.4575, "step": 127 }, { "epoch": 0.37925925925925924, "grad_norm": 9.9375, "learning_rate": 1.4914200822322255e-05, "loss": 1.5091, "step": 128 }, { "epoch": 0.38222222222222224, "grad_norm": 12.5, "learning_rate": 1.482792202730745e-05, "loss": 1.6976, "step": 129 }, { "epoch": 0.3851851851851852, "grad_norm": 11.625, "learning_rate": 1.4741172035613884e-05, "loss": 1.5555, "step": 130 }, { "epoch": 0.38814814814814813, "grad_norm": 10.3125, "learning_rate": 1.4653959313887813e-05, "loss": 1.7106, "step": 131 }, { "epoch": 0.39111111111111113, "grad_norm": 10.3125, "learning_rate": 1.4566292373937133e-05, "loss": 1.4944, "step": 132 }, { "epoch": 0.3940740740740741, "grad_norm": 9.3125, "learning_rate": 1.4478179771900634e-05, "loss": 1.4851, "step": 133 }, { "epoch": 0.397037037037037, "grad_norm": 9.0, "learning_rate": 1.4389630107412942e-05, "loss": 1.5191, "step": 134 }, { "epoch": 0.4, "grad_norm": 10.625, "learning_rate": 1.4300652022765207e-05, "loss": 1.3619, "step": 135 }, { "epoch": 0.40296296296296297, "grad_norm": 10.4375, "learning_rate": 1.4211254202061632e-05, "loss": 1.457, "step": 136 }, { "epoch": 0.4059259259259259, "grad_norm": 9.3125, "learning_rate": 1.4121445370371922e-05, "loss": 1.4453, "step": 137 }, { "epoch": 0.4088888888888889, "grad_norm": 9.375, "learning_rate": 1.4031234292879726e-05, "loss": 1.4961, "step": 138 }, { "epoch": 0.41185185185185186, "grad_norm": 9.625, "learning_rate": 1.394062977402717e-05, "loss": 1.4724, "step": 139 }, { "epoch": 0.4148148148148148, "grad_norm": 10.3125, "learning_rate": 1.3849640656655572e-05, "loss": 1.4337, "step": 140 }, { "epoch": 0.4177777777777778, "grad_norm": 11.1875, "learning_rate": 1.3758275821142382e-05, "loss": 1.4608, "step": 141 }, { "epoch": 0.42074074074074075, "grad_norm": 10.0, "learning_rate": 1.3666544184534483e-05, "loss": 1.4542, "step": 142 }, { "epoch": 0.4237037037037037, "grad_norm": 10.6875, "learning_rate": 1.3574454699677893e-05, "loss": 1.5512, "step": 143 }, { "epoch": 0.4266666666666667, "grad_norm": 10.3125, "learning_rate": 1.348201635434399e-05, "loss": 1.56, "step": 144 }, { "epoch": 0.42962962962962964, "grad_norm": 9.625, "learning_rate": 1.3389238170352318e-05, "loss": 1.3596, "step": 145 }, { "epoch": 0.4325925925925926, "grad_norm": 9.5, "learning_rate": 1.329612920269008e-05, "loss": 1.4438, "step": 146 }, { "epoch": 0.43555555555555553, "grad_norm": 10.1875, "learning_rate": 1.3202698538628376e-05, "loss": 1.4873, "step": 147 }, { "epoch": 0.43851851851851853, "grad_norm": 9.8125, "learning_rate": 1.3108955296835313e-05, "loss": 1.4859, "step": 148 }, { "epoch": 0.4414814814814815, "grad_norm": 9.75, "learning_rate": 1.3014908626486032e-05, "loss": 1.4505, "step": 149 }, { "epoch": 0.4444444444444444, "grad_norm": 9.9375, "learning_rate": 1.292056770636976e-05, "loss": 1.3619, "step": 150 }, { "epoch": 0.4474074074074074, "grad_norm": 10.25, "learning_rate": 1.282594174399399e-05, "loss": 1.4956, "step": 151 }, { "epoch": 0.45037037037037037, "grad_norm": 9.6875, "learning_rate": 1.2731039974685833e-05, "loss": 1.4486, "step": 152 }, { "epoch": 0.4533333333333333, "grad_norm": 9.5625, "learning_rate": 1.2635871660690677e-05, "loss": 1.377, "step": 153 }, { "epoch": 0.4562962962962963, "grad_norm": 9.4375, "learning_rate": 1.2540446090268193e-05, "loss": 1.4245, "step": 154 }, { "epoch": 0.45925925925925926, "grad_norm": 9.375, "learning_rate": 1.2444772576785828e-05, "loss": 1.407, "step": 155 }, { "epoch": 0.4622222222222222, "grad_norm": 9.5625, "learning_rate": 1.234886045780984e-05, "loss": 1.4341, "step": 156 }, { "epoch": 0.4651851851851852, "grad_norm": 9.3125, "learning_rate": 1.225271909419395e-05, "loss": 1.5296, "step": 157 }, { "epoch": 0.46814814814814815, "grad_norm": 9.0625, "learning_rate": 1.2156357869165771e-05, "loss": 1.4201, "step": 158 }, { "epoch": 0.4711111111111111, "grad_norm": 9.0625, "learning_rate": 1.2059786187410984e-05, "loss": 1.458, "step": 159 }, { "epoch": 0.4740740740740741, "grad_norm": 9.8125, "learning_rate": 1.1963013474155487e-05, "loss": 1.5306, "step": 160 }, { "epoch": 0.47703703703703704, "grad_norm": 9.4375, "learning_rate": 1.186604917424549e-05, "loss": 1.437, "step": 161 }, { "epoch": 0.48, "grad_norm": 9.5, "learning_rate": 1.176890275122573e-05, "loss": 1.4678, "step": 162 }, { "epoch": 0.482962962962963, "grad_norm": 9.0625, "learning_rate": 1.1671583686415833e-05, "loss": 1.2917, "step": 163 }, { "epoch": 0.48592592592592593, "grad_norm": 9.25, "learning_rate": 1.1574101477984966e-05, "loss": 1.4942, "step": 164 }, { "epoch": 0.4888888888888889, "grad_norm": 9.5625, "learning_rate": 1.1476465640024814e-05, "loss": 1.458, "step": 165 }, { "epoch": 0.4918518518518519, "grad_norm": 10.0625, "learning_rate": 1.1378685701621047e-05, "loss": 1.4213, "step": 166 }, { "epoch": 0.4948148148148148, "grad_norm": 8.625, "learning_rate": 1.1280771205923269e-05, "loss": 1.2903, "step": 167 }, { "epoch": 0.49777777777777776, "grad_norm": 9.125, "learning_rate": 1.1182731709213658e-05, "loss": 1.3953, "step": 168 }, { "epoch": 0.5007407407407407, "grad_norm": 9.5, "learning_rate": 1.1084576779974257e-05, "loss": 1.4805, "step": 169 }, { "epoch": 0.5037037037037037, "grad_norm": 9.75, "learning_rate": 1.0986315997953118e-05, "loss": 1.4822, "step": 170 }, { "epoch": 0.5066666666666667, "grad_norm": 9.75, "learning_rate": 1.0887958953229349e-05, "loss": 1.4369, "step": 171 }, { "epoch": 0.5096296296296297, "grad_norm": 9.375, "learning_rate": 1.078951524527712e-05, "loss": 1.3851, "step": 172 }, { "epoch": 0.5125925925925926, "grad_norm": 10.375, "learning_rate": 1.069099448202878e-05, "loss": 1.6454, "step": 173 }, { "epoch": 0.5155555555555555, "grad_norm": 9.1875, "learning_rate": 1.0592406278937143e-05, "loss": 1.4134, "step": 174 }, { "epoch": 0.5185185185185185, "grad_norm": 9.625, "learning_rate": 1.049376025803703e-05, "loss": 1.3613, "step": 175 }, { "epoch": 0.5214814814814814, "grad_norm": 9.5625, "learning_rate": 1.039506604700618e-05, "loss": 1.3413, "step": 176 }, { "epoch": 0.5244444444444445, "grad_norm": 8.875, "learning_rate": 1.0296333278225599e-05, "loss": 1.3768, "step": 177 }, { "epoch": 0.5274074074074074, "grad_norm": 9.875, "learning_rate": 1.0197571587839466e-05, "loss": 1.4624, "step": 178 }, { "epoch": 0.5303703703703704, "grad_norm": 9.5, "learning_rate": 1.0098790614814658e-05, "loss": 1.3982, "step": 179 }, { "epoch": 0.5333333333333333, "grad_norm": 9.0, "learning_rate": 1e-05, "loss": 1.4345, "step": 180 }, { "epoch": 0.5362962962962963, "grad_norm": 9.4375, "learning_rate": 9.901209385185345e-06, "loss": 1.4967, "step": 181 }, { "epoch": 0.5392592592592592, "grad_norm": 10.0625, "learning_rate": 9.802428412160538e-06, "loss": 1.5533, "step": 182 }, { "epoch": 0.5422222222222223, "grad_norm": 10.375, "learning_rate": 9.703666721774403e-06, "loss": 1.4622, "step": 183 }, { "epoch": 0.5451851851851852, "grad_norm": 9.4375, "learning_rate": 9.604933952993822e-06, "loss": 1.4235, "step": 184 }, { "epoch": 0.5481481481481482, "grad_norm": 9.5, "learning_rate": 9.506239741962971e-06, "loss": 1.4563, "step": 185 }, { "epoch": 0.5511111111111111, "grad_norm": 9.6875, "learning_rate": 9.407593721062858e-06, "loss": 1.5149, "step": 186 }, { "epoch": 0.554074074074074, "grad_norm": 9.25, "learning_rate": 9.309005517971222e-06, "loss": 1.4514, "step": 187 }, { "epoch": 0.557037037037037, "grad_norm": 9.8125, "learning_rate": 9.210484754722882e-06, "loss": 1.4964, "step": 188 }, { "epoch": 0.56, "grad_norm": 9.375, "learning_rate": 9.112041046770653e-06, "loss": 1.3697, "step": 189 }, { "epoch": 0.562962962962963, "grad_norm": 10.25, "learning_rate": 9.013684002046883e-06, "loss": 1.2907, "step": 190 }, { "epoch": 0.5659259259259259, "grad_norm": 9.625, "learning_rate": 8.915423220025747e-06, "loss": 1.3109, "step": 191 }, { "epoch": 0.5688888888888889, "grad_norm": 9.25, "learning_rate": 8.817268290786343e-06, "loss": 1.4432, "step": 192 }, { "epoch": 0.5718518518518518, "grad_norm": 9.6875, "learning_rate": 8.719228794076733e-06, "loss": 1.4068, "step": 193 }, { "epoch": 0.5748148148148148, "grad_norm": 10.75, "learning_rate": 8.621314298378958e-06, "loss": 1.4999, "step": 194 }, { "epoch": 0.5777777777777777, "grad_norm": 9.875, "learning_rate": 8.52353435997519e-06, "loss": 1.5483, "step": 195 }, { "epoch": 0.5807407407407408, "grad_norm": 9.25, "learning_rate": 8.425898522015038e-06, "loss": 1.3818, "step": 196 }, { "epoch": 0.5837037037037037, "grad_norm": 9.8125, "learning_rate": 8.328416313584169e-06, "loss": 1.4961, "step": 197 }, { "epoch": 0.5866666666666667, "grad_norm": 9.8125, "learning_rate": 8.231097248774273e-06, "loss": 1.4898, "step": 198 }, { "epoch": 0.5896296296296296, "grad_norm": 9.4375, "learning_rate": 8.133950825754511e-06, "loss": 1.3789, "step": 199 }, { "epoch": 0.5925925925925926, "grad_norm": 9.5625, "learning_rate": 8.036986525844516e-06, "loss": 1.3425, "step": 200 }, { "epoch": 0.5955555555555555, "grad_norm": 10.4375, "learning_rate": 7.940213812589018e-06, "loss": 1.6482, "step": 201 }, { "epoch": 0.5985185185185186, "grad_norm": 9.875, "learning_rate": 7.843642130834232e-06, "loss": 1.4535, "step": 202 }, { "epoch": 0.6014814814814815, "grad_norm": 9.5625, "learning_rate": 7.747280905806051e-06, "loss": 1.3676, "step": 203 }, { "epoch": 0.6044444444444445, "grad_norm": 9.6875, "learning_rate": 7.651139542190164e-06, "loss": 1.498, "step": 204 }, { "epoch": 0.6074074074074074, "grad_norm": 9.75, "learning_rate": 7.555227423214174e-06, "loss": 1.4675, "step": 205 }, { "epoch": 0.6103703703703703, "grad_norm": 9.75, "learning_rate": 7.45955390973181e-06, "loss": 1.5049, "step": 206 }, { "epoch": 0.6133333333333333, "grad_norm": 9.125, "learning_rate": 7.364128339309326e-06, "loss": 1.4279, "step": 207 }, { "epoch": 0.6162962962962963, "grad_norm": 9.5625, "learning_rate": 7.268960025314169e-06, "loss": 1.504, "step": 208 }, { "epoch": 0.6192592592592593, "grad_norm": 9.0625, "learning_rate": 7.174058256006012e-06, "loss": 1.4354, "step": 209 }, { "epoch": 0.6222222222222222, "grad_norm": 9.1875, "learning_rate": 7.079432293630244e-06, "loss": 1.3362, "step": 210 }, { "epoch": 0.6251851851851852, "grad_norm": 9.9375, "learning_rate": 6.985091373513972e-06, "loss": 1.3616, "step": 211 }, { "epoch": 0.6281481481481481, "grad_norm": 9.4375, "learning_rate": 6.8910447031646884e-06, "loss": 1.4284, "step": 212 }, { "epoch": 0.6311111111111111, "grad_norm": 9.0, "learning_rate": 6.797301461371626e-06, "loss": 1.2881, "step": 213 }, { "epoch": 0.6340740740740741, "grad_norm": 10.0625, "learning_rate": 6.703870797309922e-06, "loss": 1.4838, "step": 214 }, { "epoch": 0.6370370370370371, "grad_norm": 10.625, "learning_rate": 6.610761829647685e-06, "loss": 1.5364, "step": 215 }, { "epoch": 0.64, "grad_norm": 9.625, "learning_rate": 6.517983645656014e-06, "loss": 1.4076, "step": 216 }, { "epoch": 0.642962962962963, "grad_norm": 9.8125, "learning_rate": 6.4255453003221115e-06, "loss": 1.3774, "step": 217 }, { "epoch": 0.6459259259259259, "grad_norm": 9.4375, "learning_rate": 6.33345581546552e-06, "loss": 1.4266, "step": 218 }, { "epoch": 0.6488888888888888, "grad_norm": 8.875, "learning_rate": 6.241724178857621e-06, "loss": 1.3279, "step": 219 }, { "epoch": 0.6518518518518519, "grad_norm": 10.5625, "learning_rate": 6.1503593433444316e-06, "loss": 1.4278, "step": 220 }, { "epoch": 0.6548148148148148, "grad_norm": 9.6875, "learning_rate": 6.059370225972834e-06, "loss": 1.454, "step": 221 }, { "epoch": 0.6577777777777778, "grad_norm": 9.5, "learning_rate": 5.96876570712028e-06, "loss": 1.3883, "step": 222 }, { "epoch": 0.6607407407407407, "grad_norm": 9.5, "learning_rate": 5.878554629628081e-06, "loss": 1.3739, "step": 223 }, { "epoch": 0.6637037037037037, "grad_norm": 9.3125, "learning_rate": 5.788745797938372e-06, "loss": 1.381, "step": 224 }, { "epoch": 0.6666666666666666, "grad_norm": 9.375, "learning_rate": 5.699347977234799e-06, "loss": 1.3763, "step": 225 }, { "epoch": 0.6696296296296296, "grad_norm": 9.4375, "learning_rate": 5.610369892587064e-06, "loss": 1.2782, "step": 226 }, { "epoch": 0.6725925925925926, "grad_norm": 9.1875, "learning_rate": 5.5218202280993725e-06, "loss": 1.428, "step": 227 }, { "epoch": 0.6755555555555556, "grad_norm": 9.4375, "learning_rate": 5.43370762606287e-06, "loss": 1.4381, "step": 228 }, { "epoch": 0.6785185185185185, "grad_norm": 9.875, "learning_rate": 5.346040686112189e-06, "loss": 1.3499, "step": 229 }, { "epoch": 0.6814814814814815, "grad_norm": 9.6875, "learning_rate": 5.25882796438612e-06, "loss": 1.3794, "step": 230 }, { "epoch": 0.6844444444444444, "grad_norm": 9.6875, "learning_rate": 5.172077972692553e-06, "loss": 1.3564, "step": 231 }, { "epoch": 0.6874074074074074, "grad_norm": 10.0, "learning_rate": 5.08579917767775e-06, "loss": 1.4274, "step": 232 }, { "epoch": 0.6903703703703704, "grad_norm": 8.9375, "learning_rate": 5.000000000000003e-06, "loss": 1.267, "step": 233 }, { "epoch": 0.6933333333333334, "grad_norm": 9.375, "learning_rate": 4.914688813507798e-06, "loss": 1.3242, "step": 234 }, { "epoch": 0.6962962962962963, "grad_norm": 9.25, "learning_rate": 4.829873944422544e-06, "loss": 1.3488, "step": 235 }, { "epoch": 0.6992592592592592, "grad_norm": 9.125, "learning_rate": 4.745563670525942e-06, "loss": 1.3237, "step": 236 }, { "epoch": 0.7022222222222222, "grad_norm": 9.5625, "learning_rate": 4.661766220352098e-06, "loss": 1.3183, "step": 237 }, { "epoch": 0.7051851851851851, "grad_norm": 9.5625, "learning_rate": 4.578489772384415e-06, "loss": 1.4587, "step": 238 }, { "epoch": 0.7081481481481482, "grad_norm": 9.5, "learning_rate": 4.495742454257418e-06, "loss": 1.2974, "step": 239 }, { "epoch": 0.7111111111111111, "grad_norm": 9.5625, "learning_rate": 4.413532341963477e-06, "loss": 1.4561, "step": 240 }, { "epoch": 0.7140740740740741, "grad_norm": 9.625, "learning_rate": 4.331867459064623e-06, "loss": 1.4047, "step": 241 }, { "epoch": 0.717037037037037, "grad_norm": 9.3125, "learning_rate": 4.250755775909465e-06, "loss": 1.4223, "step": 242 }, { "epoch": 0.72, "grad_norm": 9.3125, "learning_rate": 4.170205208855281e-06, "loss": 1.4031, "step": 243 }, { "epoch": 0.7229629629629629, "grad_norm": 9.125, "learning_rate": 4.090223619495419e-06, "loss": 1.3283, "step": 244 }, { "epoch": 0.725925925925926, "grad_norm": 9.375, "learning_rate": 4.010818813892e-06, "loss": 1.419, "step": 245 }, { "epoch": 0.7288888888888889, "grad_norm": 9.25, "learning_rate": 3.931998541814069e-06, "loss": 1.3429, "step": 246 }, { "epoch": 0.7318518518518519, "grad_norm": 9.3125, "learning_rate": 3.85377049598123e-06, "loss": 1.436, "step": 247 }, { "epoch": 0.7348148148148148, "grad_norm": 9.375, "learning_rate": 3.7761423113128427e-06, "loss": 1.3918, "step": 248 }, { "epoch": 0.7377777777777778, "grad_norm": 9.6875, "learning_rate": 3.6991215641828903e-06, "loss": 1.3523, "step": 249 }, { "epoch": 0.7407407407407407, "grad_norm": 9.4375, "learning_rate": 3.622715771680508e-06, "loss": 1.3808, "step": 250 }, { "epoch": 0.7437037037037038, "grad_norm": 9.375, "learning_rate": 3.5469323908763507e-06, "loss": 1.4308, "step": 251 }, { "epoch": 0.7466666666666667, "grad_norm": 9.4375, "learning_rate": 3.4717788180947855e-06, "loss": 1.3985, "step": 252 }, { "epoch": 0.7496296296296296, "grad_norm": 9.625, "learning_rate": 3.3972623881920296e-06, "loss": 1.3034, "step": 253 }, { "epoch": 0.7525925925925926, "grad_norm": 9.3125, "learning_rate": 3.323390373840276e-06, "loss": 1.3579, "step": 254 }, { "epoch": 0.7555555555555555, "grad_norm": 9.25, "learning_rate": 3.250169984817897e-06, "loss": 1.3301, "step": 255 }, { "epoch": 0.7585185185185185, "grad_norm": 9.125, "learning_rate": 3.1776083673057834e-06, "loss": 1.308, "step": 256 }, { "epoch": 0.7614814814814815, "grad_norm": 9.1875, "learning_rate": 3.1057126031898843e-06, "loss": 1.3278, "step": 257 }, { "epoch": 0.7644444444444445, "grad_norm": 9.4375, "learning_rate": 3.0344897093700333e-06, "loss": 1.3678, "step": 258 }, { "epoch": 0.7674074074074074, "grad_norm": 9.25, "learning_rate": 2.963946637075107e-06, "loss": 1.2888, "step": 259 }, { "epoch": 0.7703703703703704, "grad_norm": 9.1875, "learning_rate": 2.8940902711846052e-06, "loss": 1.2899, "step": 260 }, { "epoch": 0.7733333333333333, "grad_norm": 8.625, "learning_rate": 2.8249274295566863e-06, "loss": 1.2525, "step": 261 }, { "epoch": 0.7762962962962963, "grad_norm": 8.8125, "learning_rate": 2.756464862362772e-06, "loss": 1.2665, "step": 262 }, { "epoch": 0.7792592592592592, "grad_norm": 9.4375, "learning_rate": 2.688709251428725e-06, "loss": 1.363, "step": 263 }, { "epoch": 0.7822222222222223, "grad_norm": 9.0625, "learning_rate": 2.6216672095827267e-06, "loss": 1.3086, "step": 264 }, { "epoch": 0.7851851851851852, "grad_norm": 9.3125, "learning_rate": 2.555345280009872e-06, "loss": 1.4259, "step": 265 }, { "epoch": 0.7881481481481482, "grad_norm": 9.125, "learning_rate": 2.4897499356135712e-06, "loss": 1.2748, "step": 266 }, { "epoch": 0.7911111111111111, "grad_norm": 9.25, "learning_rate": 2.424887578383799e-06, "loss": 1.329, "step": 267 }, { "epoch": 0.794074074074074, "grad_norm": 9.5625, "learning_rate": 2.3607645387722753e-06, "loss": 1.4323, "step": 268 }, { "epoch": 0.797037037037037, "grad_norm": 9.5625, "learning_rate": 2.2973870750746253e-06, "loss": 1.3081, "step": 269 }, { "epoch": 0.8, "grad_norm": 9.375, "learning_rate": 2.234761372819577e-06, "loss": 1.3967, "step": 270 }, { "epoch": 0.802962962962963, "grad_norm": 9.3125, "learning_rate": 2.1728935441652687e-06, "loss": 1.345, "step": 271 }, { "epoch": 0.8059259259259259, "grad_norm": 9.0, "learning_rate": 2.111789627302707e-06, "loss": 1.3009, "step": 272 }, { "epoch": 0.8088888888888889, "grad_norm": 9.25, "learning_rate": 2.0514555858664663e-06, "loss": 1.3242, "step": 273 }, { "epoch": 0.8118518518518518, "grad_norm": 9.3125, "learning_rate": 1.991897308352624e-06, "loss": 1.3109, "step": 274 }, { "epoch": 0.8148148148148148, "grad_norm": 8.9375, "learning_rate": 1.93312060754407e-06, "loss": 1.3353, "step": 275 }, { "epoch": 0.8177777777777778, "grad_norm": 9.625, "learning_rate": 1.875131219943187e-06, "loss": 1.2318, "step": 276 }, { "epoch": 0.8207407407407408, "grad_norm": 9.0, "learning_rate": 1.817934805211976e-06, "loss": 1.2841, "step": 277 }, { "epoch": 0.8237037037037037, "grad_norm": 8.6875, "learning_rate": 1.761536945619684e-06, "loss": 1.2261, "step": 278 }, { "epoch": 0.8266666666666667, "grad_norm": 9.5625, "learning_rate": 1.7059431454979825e-06, "loss": 1.364, "step": 279 }, { "epoch": 0.8296296296296296, "grad_norm": 9.125, "learning_rate": 1.6511588307037596e-06, "loss": 1.426, "step": 280 }, { "epoch": 0.8325925925925926, "grad_norm": 9.625, "learning_rate": 1.5971893480895583e-06, "loss": 1.458, "step": 281 }, { "epoch": 0.8355555555555556, "grad_norm": 8.875, "learning_rate": 1.5440399649817384e-06, "loss": 1.4178, "step": 282 }, { "epoch": 0.8385185185185186, "grad_norm": 9.0625, "learning_rate": 1.4917158686663992e-06, "loss": 1.3404, "step": 283 }, { "epoch": 0.8414814814814815, "grad_norm": 8.75, "learning_rate": 1.4402221658830963e-06, "loss": 1.3145, "step": 284 }, { "epoch": 0.8444444444444444, "grad_norm": 9.3125, "learning_rate": 1.3895638823264447e-06, "loss": 1.2529, "step": 285 }, { "epoch": 0.8474074074074074, "grad_norm": 10.6875, "learning_rate": 1.339745962155613e-06, "loss": 1.5298, "step": 286 }, { "epoch": 0.8503703703703703, "grad_norm": 8.75, "learning_rate": 1.2907732675117878e-06, "loss": 1.2976, "step": 287 }, { "epoch": 0.8533333333333334, "grad_norm": 9.875, "learning_rate": 1.2426505780436326e-06, "loss": 1.3695, "step": 288 }, { "epoch": 0.8562962962962963, "grad_norm": 8.75, "learning_rate": 1.1953825904408033e-06, "loss": 1.2496, "step": 289 }, { "epoch": 0.8592592592592593, "grad_norm": 9.8125, "learning_rate": 1.1489739179755622e-06, "loss": 1.4167, "step": 290 }, { "epoch": 0.8622222222222222, "grad_norm": 9.5, "learning_rate": 1.1034290900525279e-06, "loss": 1.3485, "step": 291 }, { "epoch": 0.8651851851851852, "grad_norm": 8.9375, "learning_rate": 1.0587525517666142e-06, "loss": 1.2987, "step": 292 }, { "epoch": 0.8681481481481481, "grad_norm": 9.3125, "learning_rate": 1.0149486634692019e-06, "loss": 1.2708, "step": 293 }, { "epoch": 0.8711111111111111, "grad_norm": 9.0625, "learning_rate": 9.720217003425648e-07, "loss": 1.405, "step": 294 }, { "epoch": 0.8740740740740741, "grad_norm": 9.125, "learning_rate": 9.299758519826274e-07, "loss": 1.3643, "step": 295 }, { "epoch": 0.8770370370370371, "grad_norm": 9.125, "learning_rate": 8.888152219900692e-07, "loss": 1.33, "step": 296 }, { "epoch": 0.88, "grad_norm": 9.6875, "learning_rate": 8.485438275698154e-07, "loss": 1.4734, "step": 297 }, { "epoch": 0.882962962962963, "grad_norm": 9.75, "learning_rate": 8.091655991389668e-07, "loss": 1.3231, "step": 298 }, { "epoch": 0.8859259259259259, "grad_norm": 8.9375, "learning_rate": 7.706843799431985e-07, "loss": 1.2256, "step": 299 }, { "epoch": 0.8888888888888888, "grad_norm": 9.6875, "learning_rate": 7.331039256816664e-07, "loss": 1.3831, "step": 300 }, { "epoch": 0.8918518518518519, "grad_norm": 8.9375, "learning_rate": 6.964279041404553e-07, "loss": 1.2743, "step": 301 }, { "epoch": 0.8948148148148148, "grad_norm": 9.1875, "learning_rate": 6.606598948346132e-07, "loss": 1.3141, "step": 302 }, { "epoch": 0.8977777777777778, "grad_norm": 9.8125, "learning_rate": 6.258033886587911e-07, "loss": 1.4539, "step": 303 }, { "epoch": 0.9007407407407407, "grad_norm": 10.3125, "learning_rate": 5.918617875465449e-07, "loss": 1.4409, "step": 304 }, { "epoch": 0.9037037037037037, "grad_norm": 9.9375, "learning_rate": 5.588384041383089e-07, "loss": 1.3793, "step": 305 }, { "epoch": 0.9066666666666666, "grad_norm": 9.8125, "learning_rate": 5.267364614580861e-07, "loss": 1.391, "step": 306 }, { "epoch": 0.9096296296296297, "grad_norm": 8.9375, "learning_rate": 4.955590925988896e-07, "loss": 1.2844, "step": 307 }, { "epoch": 0.9125925925925926, "grad_norm": 9.6875, "learning_rate": 4.653093404169573e-07, "loss": 1.3854, "step": 308 }, { "epoch": 0.9155555555555556, "grad_norm": 9.9375, "learning_rate": 4.359901572347758e-07, "loss": 1.378, "step": 309 }, { "epoch": 0.9185185185185185, "grad_norm": 10.0625, "learning_rate": 4.0760440455293703e-07, "loss": 1.4957, "step": 310 }, { "epoch": 0.9214814814814815, "grad_norm": 10.0625, "learning_rate": 3.801548527708621e-07, "loss": 1.518, "step": 311 }, { "epoch": 0.9244444444444444, "grad_norm": 9.1875, "learning_rate": 3.5364418091641374e-07, "loss": 1.2608, "step": 312 }, { "epoch": 0.9274074074074075, "grad_norm": 9.5625, "learning_rate": 3.280749763844293e-07, "loss": 1.3896, "step": 313 }, { "epoch": 0.9303703703703704, "grad_norm": 9.4375, "learning_rate": 3.034497346841958e-07, "loss": 1.4369, "step": 314 }, { "epoch": 0.9333333333333333, "grad_norm": 9.0, "learning_rate": 2.7977085919589253e-07, "loss": 1.4707, "step": 315 }, { "epoch": 0.9362962962962963, "grad_norm": 8.875, "learning_rate": 2.570406609360221e-07, "loss": 1.2772, "step": 316 }, { "epoch": 0.9392592592592592, "grad_norm": 9.6875, "learning_rate": 2.3526135833186527e-07, "loss": 1.3784, "step": 317 }, { "epoch": 0.9422222222222222, "grad_norm": 9.3125, "learning_rate": 2.1443507700495968e-07, "loss": 1.3823, "step": 318 }, { "epoch": 0.9451851851851852, "grad_norm": 9.375, "learning_rate": 1.9456384956365149e-07, "loss": 1.3478, "step": 319 }, { "epoch": 0.9481481481481482, "grad_norm": 9.875, "learning_rate": 1.7564961540470492e-07, "loss": 1.3708, "step": 320 }, { "epoch": 0.9511111111111111, "grad_norm": 9.125, "learning_rate": 1.5769422052403172e-07, "loss": 1.2729, "step": 321 }, { "epoch": 0.9540740740740741, "grad_norm": 9.3125, "learning_rate": 1.4069941733651948e-07, "loss": 1.3488, "step": 322 }, { "epoch": 0.957037037037037, "grad_norm": 9.0, "learning_rate": 1.2466686450499866e-07, "loss": 1.3308, "step": 323 }, { "epoch": 0.96, "grad_norm": 9.1875, "learning_rate": 1.0959812677835968e-07, "loss": 1.3702, "step": 324 }, { "epoch": 0.9629629629629629, "grad_norm": 9.125, "learning_rate": 9.549467483884412e-08, "loss": 1.2894, "step": 325 }, { "epoch": 0.965925925925926, "grad_norm": 9.6875, "learning_rate": 8.235788515849607e-08, "loss": 1.4597, "step": 326 }, { "epoch": 0.9688888888888889, "grad_norm": 8.9375, "learning_rate": 7.018903986483083e-08, "loss": 1.3452, "step": 327 }, { "epoch": 0.9718518518518519, "grad_norm": 9.1875, "learning_rate": 5.89893266156949e-08, "loss": 1.3317, "step": 328 }, { "epoch": 0.9748148148148148, "grad_norm": 9.9375, "learning_rate": 4.8759838483358745e-08, "loss": 1.325, "step": 329 }, { "epoch": 0.9777777777777777, "grad_norm": 9.125, "learning_rate": 3.950157384783104e-08, "loss": 1.3164, "step": 330 }, { "epoch": 0.9807407407407407, "grad_norm": 9.4375, "learning_rate": 3.12154362994177e-08, "loss": 1.4195, "step": 331 }, { "epoch": 0.9837037037037037, "grad_norm": 9.375, "learning_rate": 2.3902234550536862e-08, "loss": 1.3961, "step": 332 }, { "epoch": 0.9866666666666667, "grad_norm": 9.1875, "learning_rate": 1.7562682356786488e-08, "loss": 1.2901, "step": 333 }, { "epoch": 0.9896296296296296, "grad_norm": 10.0, "learning_rate": 1.2197398447283404e-08, "loss": 1.5034, "step": 334 }, { "epoch": 0.9925925925925926, "grad_norm": 10.0625, "learning_rate": 7.806906464281617e-09, "loss": 1.4242, "step": 335 }, { "epoch": 0.9955555555555555, "grad_norm": 9.625, "learning_rate": 4.39163491205652e-09, "loss": 1.4248, "step": 336 }, { "epoch": 0.9985185185185185, "grad_norm": 9.8125, "learning_rate": 1.951917115091684e-09, "loss": 1.3997, "step": 337 }, { "epoch": 1.0, "grad_norm": 13.5625, "learning_rate": 4.879911855426578e-10, "loss": 1.3577, "step": 338 } ], "logging_steps": 1, "max_steps": 338, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5614551709925376.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }