| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.956190476190476, |
| "eval_steps": 500, |
| "global_step": 325, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.015238095238095238, |
| "grad_norm": 6.41825703019793, |
| "learning_rate": 2.4242424242424244e-06, |
| "loss": 1.0212, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.030476190476190476, |
| "grad_norm": 6.32460329298527, |
| "learning_rate": 4.848484848484849e-06, |
| "loss": 1.0078, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.045714285714285714, |
| "grad_norm": 5.823493836887707, |
| "learning_rate": 7.272727272727273e-06, |
| "loss": 0.9926, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.06095238095238095, |
| "grad_norm": 4.215092249626398, |
| "learning_rate": 9.696969696969698e-06, |
| "loss": 0.9511, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0761904761904762, |
| "grad_norm": 2.755821783551069, |
| "learning_rate": 1.2121212121212122e-05, |
| "loss": 0.924, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.09142857142857143, |
| "grad_norm": 5.79854862437273, |
| "learning_rate": 1.4545454545454546e-05, |
| "loss": 0.9436, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.10666666666666667, |
| "grad_norm": 6.970153409745764, |
| "learning_rate": 1.6969696969696972e-05, |
| "loss": 0.9199, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.1219047619047619, |
| "grad_norm": 6.398045150850651, |
| "learning_rate": 1.9393939393939395e-05, |
| "loss": 0.9368, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.13714285714285715, |
| "grad_norm": 3.7323118823746717, |
| "learning_rate": 2.1818181818181818e-05, |
| "loss": 0.8624, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.1523809523809524, |
| "grad_norm": 2.4649345087328935, |
| "learning_rate": 2.4242424242424244e-05, |
| "loss": 0.8177, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.1676190476190476, |
| "grad_norm": 2.624350847053939, |
| "learning_rate": 2.6666666666666667e-05, |
| "loss": 0.7966, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.18285714285714286, |
| "grad_norm": 1.495708592085656, |
| "learning_rate": 2.9090909090909093e-05, |
| "loss": 0.7813, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.1980952380952381, |
| "grad_norm": 1.4871227985381805, |
| "learning_rate": 3.151515151515152e-05, |
| "loss": 0.7453, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.21333333333333335, |
| "grad_norm": 2.060287535099846, |
| "learning_rate": 3.3939393939393945e-05, |
| "loss": 0.7371, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.22857142857142856, |
| "grad_norm": 1.4326857447271233, |
| "learning_rate": 3.6363636363636364e-05, |
| "loss": 0.7234, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.2438095238095238, |
| "grad_norm": 2.1165784341208105, |
| "learning_rate": 3.878787878787879e-05, |
| "loss": 0.7143, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.259047619047619, |
| "grad_norm": 1.7376117944761271, |
| "learning_rate": 4.1212121212121216e-05, |
| "loss": 0.7091, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.2742857142857143, |
| "grad_norm": 1.9235494481574669, |
| "learning_rate": 4.3636363636363636e-05, |
| "loss": 0.7033, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.2895238095238095, |
| "grad_norm": 1.2885542410242774, |
| "learning_rate": 4.606060606060607e-05, |
| "loss": 0.693, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.3047619047619048, |
| "grad_norm": 2.0682542365556746, |
| "learning_rate": 4.848484848484849e-05, |
| "loss": 0.6782, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 1.1294865223043438, |
| "learning_rate": 5.0909090909090914e-05, |
| "loss": 0.6734, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.3352380952380952, |
| "grad_norm": 1.6892295840685134, |
| "learning_rate": 5.333333333333333e-05, |
| "loss": 0.6701, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.3504761904761905, |
| "grad_norm": 1.7015145087043497, |
| "learning_rate": 5.5757575757575766e-05, |
| "loss": 0.6747, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.3657142857142857, |
| "grad_norm": 1.105513245543281, |
| "learning_rate": 5.8181818181818185e-05, |
| "loss": 0.6569, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.38095238095238093, |
| "grad_norm": 1.9301412200235886, |
| "learning_rate": 6.060606060606061e-05, |
| "loss": 0.6672, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.3961904761904762, |
| "grad_norm": 1.5372986311163992, |
| "learning_rate": 6.303030303030304e-05, |
| "loss": 0.6575, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.4114285714285714, |
| "grad_norm": 1.841770844497413, |
| "learning_rate": 6.545454545454546e-05, |
| "loss": 0.6439, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.4266666666666667, |
| "grad_norm": 1.285994127925713, |
| "learning_rate": 6.787878787878789e-05, |
| "loss": 0.6421, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.4419047619047619, |
| "grad_norm": 1.306641316576376, |
| "learning_rate": 7.03030303030303e-05, |
| "loss": 0.6446, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.45714285714285713, |
| "grad_norm": 2.060289337588267, |
| "learning_rate": 7.272727272727273e-05, |
| "loss": 0.6418, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.4723809523809524, |
| "grad_norm": 1.5995121846928928, |
| "learning_rate": 7.515151515151517e-05, |
| "loss": 0.6318, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.4876190476190476, |
| "grad_norm": 1.5361156868763157, |
| "learning_rate": 7.757575757575758e-05, |
| "loss": 0.6356, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.5028571428571429, |
| "grad_norm": 1.9790958633956652, |
| "learning_rate": 8e-05, |
| "loss": 0.6315, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.518095238095238, |
| "grad_norm": 1.0983212679170282, |
| "learning_rate": 7.999768495280586e-05, |
| "loss": 0.6212, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.5333333333333333, |
| "grad_norm": 2.356052428059575, |
| "learning_rate": 7.999074007919565e-05, |
| "loss": 0.6366, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.5485714285714286, |
| "grad_norm": 1.9118106657858278, |
| "learning_rate": 7.997916618305483e-05, |
| "loss": 0.6337, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.5638095238095238, |
| "grad_norm": 2.086875421088417, |
| "learning_rate": 7.996296460408921e-05, |
| "loss": 0.6243, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.579047619047619, |
| "grad_norm": 1.6815412319889804, |
| "learning_rate": 7.994213721766979e-05, |
| "loss": 0.6222, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.5942857142857143, |
| "grad_norm": 1.9619509595749545, |
| "learning_rate": 7.99166864346157e-05, |
| "loss": 0.6185, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.6095238095238096, |
| "grad_norm": 1.1850043703613138, |
| "learning_rate": 7.988661520091513e-05, |
| "loss": 0.6222, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.6247619047619047, |
| "grad_norm": 1.5413225249585591, |
| "learning_rate": 7.985192699738432e-05, |
| "loss": 0.6131, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 2.2994933443512697, |
| "learning_rate": 7.981262583926472e-05, |
| "loss": 0.6326, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.6552380952380953, |
| "grad_norm": 0.8610367129145644, |
| "learning_rate": 7.976871627575808e-05, |
| "loss": 0.6057, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.6704761904761904, |
| "grad_norm": 2.630926210907478, |
| "learning_rate": 7.972020338950004e-05, |
| "loss": 0.6366, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.6857142857142857, |
| "grad_norm": 1.542557855004854, |
| "learning_rate": 7.96670927959716e-05, |
| "loss": 0.624, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.700952380952381, |
| "grad_norm": 1.9643148826366752, |
| "learning_rate": 7.960939064284934e-05, |
| "loss": 0.619, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.7161904761904762, |
| "grad_norm": 1.248523269079334, |
| "learning_rate": 7.954710360929362e-05, |
| "loss": 0.6108, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.7314285714285714, |
| "grad_norm": 1.225316205009982, |
| "learning_rate": 7.948023890517557e-05, |
| "loss": 0.6126, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.7466666666666667, |
| "grad_norm": 1.4200551325493893, |
| "learning_rate": 7.940880427024243e-05, |
| "loss": 0.6093, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.7619047619047619, |
| "grad_norm": 1.3484644010507392, |
| "learning_rate": 7.933280797322181e-05, |
| "loss": 0.5995, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.7771428571428571, |
| "grad_norm": 1.317639647361387, |
| "learning_rate": 7.925225881086437e-05, |
| "loss": 0.5967, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.7923809523809524, |
| "grad_norm": 1.5340582175990225, |
| "learning_rate": 7.916716610692578e-05, |
| "loss": 0.5981, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.8076190476190476, |
| "grad_norm": 1.0386828999827622, |
| "learning_rate": 7.907753971108728e-05, |
| "loss": 0.5896, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.8228571428571428, |
| "grad_norm": 1.5332648656890953, |
| "learning_rate": 7.898338999781567e-05, |
| "loss": 0.589, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.8380952380952381, |
| "grad_norm": 1.1339766871931323, |
| "learning_rate": 7.888472786516246e-05, |
| "loss": 0.5937, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.8533333333333334, |
| "grad_norm": 1.5164034405411189, |
| "learning_rate": 7.87815647335023e-05, |
| "loss": 0.5883, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.8685714285714285, |
| "grad_norm": 0.9700436403678562, |
| "learning_rate": 7.86739125442111e-05, |
| "loss": 0.5931, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.8838095238095238, |
| "grad_norm": 1.3463613799484717, |
| "learning_rate": 7.856178375828384e-05, |
| "loss": 0.589, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.8990476190476191, |
| "grad_norm": 0.8752856535881799, |
| "learning_rate": 7.844519135489204e-05, |
| "loss": 0.5754, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.9142857142857143, |
| "grad_norm": 1.1854440512880053, |
| "learning_rate": 7.832414882988153e-05, |
| "loss": 0.5894, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.9295238095238095, |
| "grad_norm": 1.7051861411911757, |
| "learning_rate": 7.819867019421023e-05, |
| "loss": 0.5885, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.9447619047619048, |
| "grad_norm": 0.6380913241526577, |
| "learning_rate": 7.806876997232625e-05, |
| "loss": 0.5834, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 1.3953430391233206, |
| "learning_rate": 7.793446320048687e-05, |
| "loss": 0.5783, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.9752380952380952, |
| "grad_norm": 1.0079170376946163, |
| "learning_rate": 7.779576542501781e-05, |
| "loss": 0.5839, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.9904761904761905, |
| "grad_norm": 1.1878161560623453, |
| "learning_rate": 7.765269270051389e-05, |
| "loss": 0.575, |
| "step": 65 |
| }, |
| { |
| "epoch": 1.0066666666666666, |
| "grad_norm": 1.495861192796852, |
| "learning_rate": 7.750526158798056e-05, |
| "loss": 0.8303, |
| "step": 66 |
| }, |
| { |
| "epoch": 1.0219047619047619, |
| "grad_norm": 1.3683956168224771, |
| "learning_rate": 7.735348915291701e-05, |
| "loss": 0.571, |
| "step": 67 |
| }, |
| { |
| "epoch": 1.0371428571428571, |
| "grad_norm": 0.8831683489698775, |
| "learning_rate": 7.719739296334073e-05, |
| "loss": 0.5548, |
| "step": 68 |
| }, |
| { |
| "epoch": 1.0523809523809524, |
| "grad_norm": 0.9884633873935289, |
| "learning_rate": 7.7036991087754e-05, |
| "loss": 0.5619, |
| "step": 69 |
| }, |
| { |
| "epoch": 1.0676190476190477, |
| "grad_norm": 1.658491152165491, |
| "learning_rate": 7.687230209305242e-05, |
| "loss": 0.5876, |
| "step": 70 |
| }, |
| { |
| "epoch": 1.0828571428571427, |
| "grad_norm": 0.8352075307968011, |
| "learning_rate": 7.670334504237575e-05, |
| "loss": 0.5685, |
| "step": 71 |
| }, |
| { |
| "epoch": 1.098095238095238, |
| "grad_norm": 1.429719144977111, |
| "learning_rate": 7.65301394929013e-05, |
| "loss": 0.5764, |
| "step": 72 |
| }, |
| { |
| "epoch": 1.1133333333333333, |
| "grad_norm": 0.9027318597099093, |
| "learning_rate": 7.635270549358012e-05, |
| "loss": 0.5602, |
| "step": 73 |
| }, |
| { |
| "epoch": 1.1285714285714286, |
| "grad_norm": 1.0536650805805048, |
| "learning_rate": 7.617106358281634e-05, |
| "loss": 0.5765, |
| "step": 74 |
| }, |
| { |
| "epoch": 1.1438095238095238, |
| "grad_norm": 0.8336607908560256, |
| "learning_rate": 7.598523478608974e-05, |
| "loss": 0.5583, |
| "step": 75 |
| }, |
| { |
| "epoch": 1.159047619047619, |
| "grad_norm": 0.7166113488943966, |
| "learning_rate": 7.579524061352206e-05, |
| "loss": 0.551, |
| "step": 76 |
| }, |
| { |
| "epoch": 1.1742857142857144, |
| "grad_norm": 0.8105369080449557, |
| "learning_rate": 7.560110305738708e-05, |
| "loss": 0.5659, |
| "step": 77 |
| }, |
| { |
| "epoch": 1.1895238095238094, |
| "grad_norm": 0.6283210931425559, |
| "learning_rate": 7.540284458956503e-05, |
| "loss": 0.5379, |
| "step": 78 |
| }, |
| { |
| "epoch": 1.2047619047619047, |
| "grad_norm": 0.8126105033449552, |
| "learning_rate": 7.520048815894144e-05, |
| "loss": 0.5476, |
| "step": 79 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.5252506521520861, |
| "learning_rate": 7.49940571887506e-05, |
| "loss": 0.5551, |
| "step": 80 |
| }, |
| { |
| "epoch": 1.2352380952380952, |
| "grad_norm": 0.6609919242663752, |
| "learning_rate": 7.478357557386444e-05, |
| "loss": 0.5407, |
| "step": 81 |
| }, |
| { |
| "epoch": 1.2504761904761905, |
| "grad_norm": 0.576238102776681, |
| "learning_rate": 7.456906767802658e-05, |
| "loss": 0.5465, |
| "step": 82 |
| }, |
| { |
| "epoch": 1.2657142857142858, |
| "grad_norm": 0.5233988001346253, |
| "learning_rate": 7.435055833103213e-05, |
| "loss": 0.55, |
| "step": 83 |
| }, |
| { |
| "epoch": 1.2809523809523808, |
| "grad_norm": 0.44023553757218015, |
| "learning_rate": 7.412807282585362e-05, |
| "loss": 0.5412, |
| "step": 84 |
| }, |
| { |
| "epoch": 1.2961904761904761, |
| "grad_norm": 0.4831178702636027, |
| "learning_rate": 7.390163691571325e-05, |
| "loss": 0.5398, |
| "step": 85 |
| }, |
| { |
| "epoch": 1.3114285714285714, |
| "grad_norm": 0.454634628403172, |
| "learning_rate": 7.367127681110199e-05, |
| "loss": 0.534, |
| "step": 86 |
| }, |
| { |
| "epoch": 1.3266666666666667, |
| "grad_norm": 0.44419681534445155, |
| "learning_rate": 7.343701917674547e-05, |
| "loss": 0.5421, |
| "step": 87 |
| }, |
| { |
| "epoch": 1.341904761904762, |
| "grad_norm": 0.4888479476988804, |
| "learning_rate": 7.319889112851771e-05, |
| "loss": 0.5371, |
| "step": 88 |
| }, |
| { |
| "epoch": 1.3571428571428572, |
| "grad_norm": 0.7143397739870562, |
| "learning_rate": 7.295692023030217e-05, |
| "loss": 0.5377, |
| "step": 89 |
| }, |
| { |
| "epoch": 1.3723809523809525, |
| "grad_norm": 0.9465106459896011, |
| "learning_rate": 7.271113449080131e-05, |
| "loss": 0.5479, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.3876190476190478, |
| "grad_norm": 1.2678262246780991, |
| "learning_rate": 7.246156236029445e-05, |
| "loss": 0.5452, |
| "step": 91 |
| }, |
| { |
| "epoch": 1.4028571428571428, |
| "grad_norm": 0.6281248994163343, |
| "learning_rate": 7.220823272734461e-05, |
| "loss": 0.5401, |
| "step": 92 |
| }, |
| { |
| "epoch": 1.418095238095238, |
| "grad_norm": 0.34247752257094766, |
| "learning_rate": 7.19511749154546e-05, |
| "loss": 0.5394, |
| "step": 93 |
| }, |
| { |
| "epoch": 1.4333333333333333, |
| "grad_norm": 0.7014077226631056, |
| "learning_rate": 7.169041867967272e-05, |
| "loss": 0.5407, |
| "step": 94 |
| }, |
| { |
| "epoch": 1.4485714285714286, |
| "grad_norm": 0.9221215435654039, |
| "learning_rate": 7.142599420314857e-05, |
| "loss": 0.5464, |
| "step": 95 |
| }, |
| { |
| "epoch": 1.4638095238095237, |
| "grad_norm": 0.8724875653280793, |
| "learning_rate": 7.115793209363929e-05, |
| "loss": 0.5379, |
| "step": 96 |
| }, |
| { |
| "epoch": 1.479047619047619, |
| "grad_norm": 0.734302784896464, |
| "learning_rate": 7.088626337996657e-05, |
| "loss": 0.5248, |
| "step": 97 |
| }, |
| { |
| "epoch": 1.4942857142857142, |
| "grad_norm": 0.6804548302488894, |
| "learning_rate": 7.06110195084251e-05, |
| "loss": 0.5403, |
| "step": 98 |
| }, |
| { |
| "epoch": 1.5095238095238095, |
| "grad_norm": 0.8180102079360684, |
| "learning_rate": 7.03322323391425e-05, |
| "loss": 0.5356, |
| "step": 99 |
| }, |
| { |
| "epoch": 1.5247619047619048, |
| "grad_norm": 0.7089574204038593, |
| "learning_rate": 7.004993414239147e-05, |
| "loss": 0.5367, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.4534978500691297, |
| "learning_rate": 6.976415759485443e-05, |
| "loss": 0.5301, |
| "step": 101 |
| }, |
| { |
| "epoch": 1.5552380952380953, |
| "grad_norm": 0.5791842384826449, |
| "learning_rate": 6.947493577584111e-05, |
| "loss": 0.5338, |
| "step": 102 |
| }, |
| { |
| "epoch": 1.5704761904761906, |
| "grad_norm": 0.6373667696236239, |
| "learning_rate": 6.918230216345951e-05, |
| "loss": 0.5255, |
| "step": 103 |
| }, |
| { |
| "epoch": 1.5857142857142859, |
| "grad_norm": 0.5813750117468534, |
| "learning_rate": 6.888629063074082e-05, |
| "loss": 0.5209, |
| "step": 104 |
| }, |
| { |
| "epoch": 1.600952380952381, |
| "grad_norm": 0.5731639034091425, |
| "learning_rate": 6.858693544171845e-05, |
| "loss": 0.5215, |
| "step": 105 |
| }, |
| { |
| "epoch": 1.6161904761904762, |
| "grad_norm": 0.5351450945596543, |
| "learning_rate": 6.828427124746191e-05, |
| "loss": 0.5301, |
| "step": 106 |
| }, |
| { |
| "epoch": 1.6314285714285715, |
| "grad_norm": 0.6016816279659157, |
| "learning_rate": 6.797833308206588e-05, |
| "loss": 0.529, |
| "step": 107 |
| }, |
| { |
| "epoch": 1.6466666666666665, |
| "grad_norm": 0.7659521464895882, |
| "learning_rate": 6.766915635859497e-05, |
| "loss": 0.533, |
| "step": 108 |
| }, |
| { |
| "epoch": 1.6619047619047618, |
| "grad_norm": 0.9630524913355004, |
| "learning_rate": 6.735677686498443e-05, |
| "loss": 0.5297, |
| "step": 109 |
| }, |
| { |
| "epoch": 1.677142857142857, |
| "grad_norm": 1.1492200252998914, |
| "learning_rate": 6.70412307598978e-05, |
| "loss": 0.5264, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.6923809523809523, |
| "grad_norm": 0.5857314920473611, |
| "learning_rate": 6.672255456854135e-05, |
| "loss": 0.5241, |
| "step": 111 |
| }, |
| { |
| "epoch": 1.7076190476190476, |
| "grad_norm": 0.48173626706101424, |
| "learning_rate": 6.640078517843619e-05, |
| "loss": 0.5346, |
| "step": 112 |
| }, |
| { |
| "epoch": 1.7228571428571429, |
| "grad_norm": 0.7730414410041352, |
| "learning_rate": 6.60759598351485e-05, |
| "loss": 0.5229, |
| "step": 113 |
| }, |
| { |
| "epoch": 1.7380952380952381, |
| "grad_norm": 0.6363250099986629, |
| "learning_rate": 6.57481161379783e-05, |
| "loss": 0.5243, |
| "step": 114 |
| }, |
| { |
| "epoch": 1.7533333333333334, |
| "grad_norm": 0.45624023704141803, |
| "learning_rate": 6.54172920356071e-05, |
| "loss": 0.5247, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.7685714285714287, |
| "grad_norm": 0.4715396085670591, |
| "learning_rate": 6.508352582170543e-05, |
| "loss": 0.5144, |
| "step": 116 |
| }, |
| { |
| "epoch": 1.783809523809524, |
| "grad_norm": 0.5945001964397519, |
| "learning_rate": 6.474685613050014e-05, |
| "loss": 0.5279, |
| "step": 117 |
| }, |
| { |
| "epoch": 1.799047619047619, |
| "grad_norm": 0.5588842197173406, |
| "learning_rate": 6.440732193230241e-05, |
| "loss": 0.522, |
| "step": 118 |
| }, |
| { |
| "epoch": 1.8142857142857143, |
| "grad_norm": 0.36397509002448036, |
| "learning_rate": 6.40649625289969e-05, |
| "loss": 0.5206, |
| "step": 119 |
| }, |
| { |
| "epoch": 1.8295238095238096, |
| "grad_norm": 0.5225246500798869, |
| "learning_rate": 6.371981754949241e-05, |
| "loss": 0.5233, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.8447619047619046, |
| "grad_norm": 0.6419719236551851, |
| "learning_rate": 6.337192694513474e-05, |
| "loss": 0.5265, |
| "step": 121 |
| }, |
| { |
| "epoch": 1.8599999999999999, |
| "grad_norm": 0.41672473350375433, |
| "learning_rate": 6.30213309850823e-05, |
| "loss": 0.5165, |
| "step": 122 |
| }, |
| { |
| "epoch": 1.8752380952380951, |
| "grad_norm": 0.29619919377664744, |
| "learning_rate": 6.266807025164472e-05, |
| "loss": 0.5158, |
| "step": 123 |
| }, |
| { |
| "epoch": 1.8904761904761904, |
| "grad_norm": 0.41731867813128104, |
| "learning_rate": 6.231218563558551e-05, |
| "loss": 0.518, |
| "step": 124 |
| }, |
| { |
| "epoch": 1.9057142857142857, |
| "grad_norm": 0.3648760435140768, |
| "learning_rate": 6.195371833138878e-05, |
| "loss": 0.5165, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.920952380952381, |
| "grad_norm": 0.3531467699123105, |
| "learning_rate": 6.159270983249084e-05, |
| "loss": 0.5247, |
| "step": 126 |
| }, |
| { |
| "epoch": 1.9361904761904762, |
| "grad_norm": 0.45588003744063627, |
| "learning_rate": 6.122920192647734e-05, |
| "loss": 0.5197, |
| "step": 127 |
| }, |
| { |
| "epoch": 1.9514285714285715, |
| "grad_norm": 0.5510619573466758, |
| "learning_rate": 6.086323669024616e-05, |
| "loss": 0.5256, |
| "step": 128 |
| }, |
| { |
| "epoch": 1.9666666666666668, |
| "grad_norm": 0.5646853258922581, |
| "learning_rate": 6.049485648513696e-05, |
| "loss": 0.527, |
| "step": 129 |
| }, |
| { |
| "epoch": 1.981904761904762, |
| "grad_norm": 0.4479646401937566, |
| "learning_rate": 6.012410395202774e-05, |
| "loss": 0.5234, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.997142857142857, |
| "grad_norm": 0.575986791359234, |
| "learning_rate": 5.9751022006399095e-05, |
| "loss": 0.7009, |
| "step": 131 |
| }, |
| { |
| "epoch": 2.013333333333333, |
| "grad_norm": 0.814219365805337, |
| "learning_rate": 5.9375653833366585e-05, |
| "loss": 0.5404, |
| "step": 132 |
| }, |
| { |
| "epoch": 2.0285714285714285, |
| "grad_norm": 0.8146199046124171, |
| "learning_rate": 5.8998042882682024e-05, |
| "loss": 0.4877, |
| "step": 133 |
| }, |
| { |
| "epoch": 2.0438095238095237, |
| "grad_norm": 0.7629490780519704, |
| "learning_rate": 5.861823286370396e-05, |
| "loss": 0.4849, |
| "step": 134 |
| }, |
| { |
| "epoch": 2.059047619047619, |
| "grad_norm": 0.6570978207497692, |
| "learning_rate": 5.823626774033836e-05, |
| "loss": 0.4887, |
| "step": 135 |
| }, |
| { |
| "epoch": 2.0742857142857143, |
| "grad_norm": 0.5305731216336181, |
| "learning_rate": 5.7852191725949564e-05, |
| "loss": 0.4881, |
| "step": 136 |
| }, |
| { |
| "epoch": 2.0895238095238096, |
| "grad_norm": 0.512493977688762, |
| "learning_rate": 5.746604927824257e-05, |
| "loss": 0.4844, |
| "step": 137 |
| }, |
| { |
| "epoch": 2.104761904761905, |
| "grad_norm": 0.4828622337949408, |
| "learning_rate": 5.707788509411685e-05, |
| "loss": 0.4887, |
| "step": 138 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.4884027438000281, |
| "learning_rate": 5.668774410449268e-05, |
| "loss": 0.4908, |
| "step": 139 |
| }, |
| { |
| "epoch": 2.1352380952380954, |
| "grad_norm": 0.42776751878992564, |
| "learning_rate": 5.629567146911023e-05, |
| "loss": 0.4872, |
| "step": 140 |
| }, |
| { |
| "epoch": 2.1504761904761907, |
| "grad_norm": 0.36115257273207996, |
| "learning_rate": 5.590171257130225e-05, |
| "loss": 0.4861, |
| "step": 141 |
| }, |
| { |
| "epoch": 2.1657142857142855, |
| "grad_norm": 0.38436733504559367, |
| "learning_rate": 5.550591301274073e-05, |
| "loss": 0.4844, |
| "step": 142 |
| }, |
| { |
| "epoch": 2.1809523809523808, |
| "grad_norm": 0.3435583779900254, |
| "learning_rate": 5.5108318608158595e-05, |
| "loss": 0.4836, |
| "step": 143 |
| }, |
| { |
| "epoch": 2.196190476190476, |
| "grad_norm": 0.2857883944660014, |
| "learning_rate": 5.4708975380046356e-05, |
| "loss": 0.4796, |
| "step": 144 |
| }, |
| { |
| "epoch": 2.2114285714285713, |
| "grad_norm": 0.31512140613431383, |
| "learning_rate": 5.430792955332503e-05, |
| "loss": 0.4863, |
| "step": 145 |
| }, |
| { |
| "epoch": 2.2266666666666666, |
| "grad_norm": 0.31769228977835245, |
| "learning_rate": 5.3905227549995366e-05, |
| "loss": 0.4809, |
| "step": 146 |
| }, |
| { |
| "epoch": 2.241904761904762, |
| "grad_norm": 0.27924100773876076, |
| "learning_rate": 5.350091598376454e-05, |
| "loss": 0.4805, |
| "step": 147 |
| }, |
| { |
| "epoch": 2.257142857142857, |
| "grad_norm": 0.2662485345450648, |
| "learning_rate": 5.3095041654650384e-05, |
| "loss": 0.4831, |
| "step": 148 |
| }, |
| { |
| "epoch": 2.2723809523809524, |
| "grad_norm": 0.23461248952656025, |
| "learning_rate": 5.268765154356425e-05, |
| "loss": 0.4816, |
| "step": 149 |
| }, |
| { |
| "epoch": 2.2876190476190477, |
| "grad_norm": 0.240243576104443, |
| "learning_rate": 5.227879280687281e-05, |
| "loss": 0.4788, |
| "step": 150 |
| }, |
| { |
| "epoch": 2.302857142857143, |
| "grad_norm": 0.24392104367270967, |
| "learning_rate": 5.186851277093961e-05, |
| "loss": 0.4851, |
| "step": 151 |
| }, |
| { |
| "epoch": 2.318095238095238, |
| "grad_norm": 0.19828908319240954, |
| "learning_rate": 5.1456858926646974e-05, |
| "loss": 0.4817, |
| "step": 152 |
| }, |
| { |
| "epoch": 2.3333333333333335, |
| "grad_norm": 0.2354905746190516, |
| "learning_rate": 5.104387892389877e-05, |
| "loss": 0.4816, |
| "step": 153 |
| }, |
| { |
| "epoch": 2.3485714285714288, |
| "grad_norm": 0.24408900516772156, |
| "learning_rate": 5.062962056610478e-05, |
| "loss": 0.4833, |
| "step": 154 |
| }, |
| { |
| "epoch": 2.3638095238095236, |
| "grad_norm": 0.22588988217182931, |
| "learning_rate": 5.021413180464748e-05, |
| "loss": 0.4884, |
| "step": 155 |
| }, |
| { |
| "epoch": 2.379047619047619, |
| "grad_norm": 0.20985654308992252, |
| "learning_rate": 4.979746073333145e-05, |
| "loss": 0.4777, |
| "step": 156 |
| }, |
| { |
| "epoch": 2.394285714285714, |
| "grad_norm": 0.24110384355800776, |
| "learning_rate": 4.9379655582816396e-05, |
| "loss": 0.4867, |
| "step": 157 |
| }, |
| { |
| "epoch": 2.4095238095238094, |
| "grad_norm": 0.20401637758198887, |
| "learning_rate": 4.896076471503439e-05, |
| "loss": 0.4766, |
| "step": 158 |
| }, |
| { |
| "epoch": 2.4247619047619047, |
| "grad_norm": 0.19932449853927872, |
| "learning_rate": 4.854083661759185e-05, |
| "loss": 0.4884, |
| "step": 159 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.24533376047455172, |
| "learning_rate": 4.8119919898156934e-05, |
| "loss": 0.4834, |
| "step": 160 |
| }, |
| { |
| "epoch": 2.455238095238095, |
| "grad_norm": 0.19336383545060395, |
| "learning_rate": 4.76980632788332e-05, |
| "loss": 0.4779, |
| "step": 161 |
| }, |
| { |
| "epoch": 2.4704761904761905, |
| "grad_norm": 0.19824370981981676, |
| "learning_rate": 4.7275315590519765e-05, |
| "loss": 0.4823, |
| "step": 162 |
| }, |
| { |
| "epoch": 2.4857142857142858, |
| "grad_norm": 0.20431330071012052, |
| "learning_rate": 4.6851725767259106e-05, |
| "loss": 0.4781, |
| "step": 163 |
| }, |
| { |
| "epoch": 2.500952380952381, |
| "grad_norm": 0.2082524151757895, |
| "learning_rate": 4.642734284057283e-05, |
| "loss": 0.4755, |
| "step": 164 |
| }, |
| { |
| "epoch": 2.5161904761904763, |
| "grad_norm": 0.2129960852067495, |
| "learning_rate": 4.600221593378611e-05, |
| "loss": 0.4796, |
| "step": 165 |
| }, |
| { |
| "epoch": 2.5314285714285716, |
| "grad_norm": 0.17147652502249622, |
| "learning_rate": 4.5576394256341596e-05, |
| "loss": 0.4784, |
| "step": 166 |
| }, |
| { |
| "epoch": 2.546666666666667, |
| "grad_norm": 0.2328806782705143, |
| "learning_rate": 4.514992709810324e-05, |
| "loss": 0.4839, |
| "step": 167 |
| }, |
| { |
| "epoch": 2.5619047619047617, |
| "grad_norm": 0.2628457128916824, |
| "learning_rate": 4.472286382365096e-05, |
| "loss": 0.4899, |
| "step": 168 |
| }, |
| { |
| "epoch": 2.5771428571428574, |
| "grad_norm": 0.20530497111629914, |
| "learning_rate": 4.429525386656653e-05, |
| "loss": 0.4865, |
| "step": 169 |
| }, |
| { |
| "epoch": 2.5923809523809522, |
| "grad_norm": 0.18615860925149122, |
| "learning_rate": 4.38671467237115e-05, |
| "loss": 0.4841, |
| "step": 170 |
| }, |
| { |
| "epoch": 2.6076190476190475, |
| "grad_norm": 0.181524499480342, |
| "learning_rate": 4.343859194949787e-05, |
| "loss": 0.4792, |
| "step": 171 |
| }, |
| { |
| "epoch": 2.6228571428571428, |
| "grad_norm": 0.2025029996461714, |
| "learning_rate": 4.3009639150152e-05, |
| "loss": 0.4788, |
| "step": 172 |
| }, |
| { |
| "epoch": 2.638095238095238, |
| "grad_norm": 0.19560730560439193, |
| "learning_rate": 4.2580337977972664e-05, |
| "loss": 0.4827, |
| "step": 173 |
| }, |
| { |
| "epoch": 2.6533333333333333, |
| "grad_norm": 0.16741577940585486, |
| "learning_rate": 4.215073812558352e-05, |
| "loss": 0.4885, |
| "step": 174 |
| }, |
| { |
| "epoch": 2.6685714285714286, |
| "grad_norm": 0.1739859347419123, |
| "learning_rate": 4.1720889320181224e-05, |
| "loss": 0.4772, |
| "step": 175 |
| }, |
| { |
| "epoch": 2.683809523809524, |
| "grad_norm": 0.19324214245630159, |
| "learning_rate": 4.1290841317779325e-05, |
| "loss": 0.4884, |
| "step": 176 |
| }, |
| { |
| "epoch": 2.699047619047619, |
| "grad_norm": 0.162133413448313, |
| "learning_rate": 4.0860643897448894e-05, |
| "loss": 0.4789, |
| "step": 177 |
| }, |
| { |
| "epoch": 2.7142857142857144, |
| "grad_norm": 0.1910560430197172, |
| "learning_rate": 4.043034685555647e-05, |
| "loss": 0.4807, |
| "step": 178 |
| }, |
| { |
| "epoch": 2.7295238095238097, |
| "grad_norm": 0.14823981314032625, |
| "learning_rate": 4e-05, |
| "loss": 0.481, |
| "step": 179 |
| }, |
| { |
| "epoch": 2.744761904761905, |
| "grad_norm": 0.1784212235327859, |
| "learning_rate": 3.9569653144443546e-05, |
| "loss": 0.481, |
| "step": 180 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.200733513756011, |
| "learning_rate": 3.913935610255112e-05, |
| "loss": 0.4763, |
| "step": 181 |
| }, |
| { |
| "epoch": 2.7752380952380955, |
| "grad_norm": 0.14938330915090647, |
| "learning_rate": 3.870915868222068e-05, |
| "loss": 0.4819, |
| "step": 182 |
| }, |
| { |
| "epoch": 2.7904761904761903, |
| "grad_norm": 0.1739261226284681, |
| "learning_rate": 3.8279110679818775e-05, |
| "loss": 0.4738, |
| "step": 183 |
| }, |
| { |
| "epoch": 2.8057142857142856, |
| "grad_norm": 0.19838890478019414, |
| "learning_rate": 3.78492618744165e-05, |
| "loss": 0.4783, |
| "step": 184 |
| }, |
| { |
| "epoch": 2.820952380952381, |
| "grad_norm": 0.1528772994448381, |
| "learning_rate": 3.7419662022027356e-05, |
| "loss": 0.4769, |
| "step": 185 |
| }, |
| { |
| "epoch": 2.836190476190476, |
| "grad_norm": 0.1919042097767604, |
| "learning_rate": 3.6990360849848005e-05, |
| "loss": 0.4757, |
| "step": 186 |
| }, |
| { |
| "epoch": 2.8514285714285714, |
| "grad_norm": 0.17566953990872838, |
| "learning_rate": 3.656140805050215e-05, |
| "loss": 0.4786, |
| "step": 187 |
| }, |
| { |
| "epoch": 2.8666666666666667, |
| "grad_norm": 0.1637970789000741, |
| "learning_rate": 3.6132853276288506e-05, |
| "loss": 0.4739, |
| "step": 188 |
| }, |
| { |
| "epoch": 2.881904761904762, |
| "grad_norm": 0.18549958217752188, |
| "learning_rate": 3.570474613343348e-05, |
| "loss": 0.4768, |
| "step": 189 |
| }, |
| { |
| "epoch": 2.8971428571428572, |
| "grad_norm": 0.18002674770058677, |
| "learning_rate": 3.527713617634904e-05, |
| "loss": 0.4733, |
| "step": 190 |
| }, |
| { |
| "epoch": 2.9123809523809525, |
| "grad_norm": 0.15898767756027013, |
| "learning_rate": 3.4850072901896766e-05, |
| "loss": 0.4802, |
| "step": 191 |
| }, |
| { |
| "epoch": 2.9276190476190473, |
| "grad_norm": 0.17974934358276728, |
| "learning_rate": 3.442360574365843e-05, |
| "loss": 0.4834, |
| "step": 192 |
| }, |
| { |
| "epoch": 2.942857142857143, |
| "grad_norm": 0.14203142815133346, |
| "learning_rate": 3.39977840662139e-05, |
| "loss": 0.4747, |
| "step": 193 |
| }, |
| { |
| "epoch": 2.958095238095238, |
| "grad_norm": 0.17952115377014885, |
| "learning_rate": 3.3572657159427186e-05, |
| "loss": 0.4808, |
| "step": 194 |
| }, |
| { |
| "epoch": 2.9733333333333336, |
| "grad_norm": 0.17353334331590287, |
| "learning_rate": 3.31482742327409e-05, |
| "loss": 0.4751, |
| "step": 195 |
| }, |
| { |
| "epoch": 2.9885714285714284, |
| "grad_norm": 0.1528037773950146, |
| "learning_rate": 3.272468440948025e-05, |
| "loss": 0.4734, |
| "step": 196 |
| }, |
| { |
| "epoch": 3.0047619047619047, |
| "grad_norm": 0.20480536463226645, |
| "learning_rate": 3.230193672116681e-05, |
| "loss": 0.6928, |
| "step": 197 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 0.19151123516717333, |
| "learning_rate": 3.188008010184306e-05, |
| "loss": 0.4452, |
| "step": 198 |
| }, |
| { |
| "epoch": 3.0352380952380953, |
| "grad_norm": 0.2104984649163227, |
| "learning_rate": 3.145916338240816e-05, |
| "loss": 0.4546, |
| "step": 199 |
| }, |
| { |
| "epoch": 3.0504761904761906, |
| "grad_norm": 0.19741714449829997, |
| "learning_rate": 3.103923528496562e-05, |
| "loss": 0.4452, |
| "step": 200 |
| }, |
| { |
| "epoch": 3.065714285714286, |
| "grad_norm": 0.19623857285807467, |
| "learning_rate": 3.062034441718362e-05, |
| "loss": 0.4483, |
| "step": 201 |
| }, |
| { |
| "epoch": 3.080952380952381, |
| "grad_norm": 0.25035024771569697, |
| "learning_rate": 3.0202539266668568e-05, |
| "loss": 0.4506, |
| "step": 202 |
| }, |
| { |
| "epoch": 3.0961904761904764, |
| "grad_norm": 0.16968910324281286, |
| "learning_rate": 2.9785868195352525e-05, |
| "loss": 0.4449, |
| "step": 203 |
| }, |
| { |
| "epoch": 3.111428571428571, |
| "grad_norm": 0.23581488026041375, |
| "learning_rate": 2.9370379433895228e-05, |
| "loss": 0.4512, |
| "step": 204 |
| }, |
| { |
| "epoch": 3.1266666666666665, |
| "grad_norm": 0.15672745746884265, |
| "learning_rate": 2.8956121076101244e-05, |
| "loss": 0.449, |
| "step": 205 |
| }, |
| { |
| "epoch": 3.1419047619047618, |
| "grad_norm": 0.20702186752340365, |
| "learning_rate": 2.8543141073353026e-05, |
| "loss": 0.4439, |
| "step": 206 |
| }, |
| { |
| "epoch": 3.157142857142857, |
| "grad_norm": 0.16133838526688454, |
| "learning_rate": 2.8131487229060387e-05, |
| "loss": 0.4427, |
| "step": 207 |
| }, |
| { |
| "epoch": 3.1723809523809523, |
| "grad_norm": 0.1854062242780086, |
| "learning_rate": 2.7721207193127212e-05, |
| "loss": 0.4472, |
| "step": 208 |
| }, |
| { |
| "epoch": 3.1876190476190476, |
| "grad_norm": 0.17182738319155882, |
| "learning_rate": 2.7312348456435766e-05, |
| "loss": 0.4453, |
| "step": 209 |
| }, |
| { |
| "epoch": 3.202857142857143, |
| "grad_norm": 0.16697225397908244, |
| "learning_rate": 2.6904958345349633e-05, |
| "loss": 0.4561, |
| "step": 210 |
| }, |
| { |
| "epoch": 3.218095238095238, |
| "grad_norm": 0.16821522604659328, |
| "learning_rate": 2.6499084016235475e-05, |
| "loss": 0.444, |
| "step": 211 |
| }, |
| { |
| "epoch": 3.2333333333333334, |
| "grad_norm": 0.14853827941880265, |
| "learning_rate": 2.609477245000464e-05, |
| "loss": 0.44, |
| "step": 212 |
| }, |
| { |
| "epoch": 3.2485714285714287, |
| "grad_norm": 0.1502037122279702, |
| "learning_rate": 2.569207044667498e-05, |
| "loss": 0.4494, |
| "step": 213 |
| }, |
| { |
| "epoch": 3.263809523809524, |
| "grad_norm": 0.16216155417715333, |
| "learning_rate": 2.529102461995364e-05, |
| "loss": 0.4471, |
| "step": 214 |
| }, |
| { |
| "epoch": 3.279047619047619, |
| "grad_norm": 0.1586744564208733, |
| "learning_rate": 2.48916813918414e-05, |
| "loss": 0.4486, |
| "step": 215 |
| }, |
| { |
| "epoch": 3.2942857142857145, |
| "grad_norm": 0.14092622353201845, |
| "learning_rate": 2.449408698725928e-05, |
| "loss": 0.445, |
| "step": 216 |
| }, |
| { |
| "epoch": 3.3095238095238093, |
| "grad_norm": 0.15562276697944483, |
| "learning_rate": 2.409828742869777e-05, |
| "loss": 0.4415, |
| "step": 217 |
| }, |
| { |
| "epoch": 3.3247619047619046, |
| "grad_norm": 0.14631955274946826, |
| "learning_rate": 2.370432853088978e-05, |
| "loss": 0.454, |
| "step": 218 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 0.14585582091880941, |
| "learning_rate": 2.3312255895507336e-05, |
| "loss": 0.4473, |
| "step": 219 |
| }, |
| { |
| "epoch": 3.355238095238095, |
| "grad_norm": 0.1458201114996886, |
| "learning_rate": 2.2922114905883167e-05, |
| "loss": 0.4444, |
| "step": 220 |
| }, |
| { |
| "epoch": 3.3704761904761904, |
| "grad_norm": 0.1402608135797642, |
| "learning_rate": 2.2533950721757447e-05, |
| "loss": 0.4465, |
| "step": 221 |
| }, |
| { |
| "epoch": 3.3857142857142857, |
| "grad_norm": 0.14672569191797688, |
| "learning_rate": 2.2147808274050436e-05, |
| "loss": 0.4484, |
| "step": 222 |
| }, |
| { |
| "epoch": 3.400952380952381, |
| "grad_norm": 0.15377591576947733, |
| "learning_rate": 2.1763732259661657e-05, |
| "loss": 0.4479, |
| "step": 223 |
| }, |
| { |
| "epoch": 3.416190476190476, |
| "grad_norm": 0.14739415052119614, |
| "learning_rate": 2.138176713629605e-05, |
| "loss": 0.4475, |
| "step": 224 |
| }, |
| { |
| "epoch": 3.4314285714285715, |
| "grad_norm": 0.14607883746358746, |
| "learning_rate": 2.1001957117318002e-05, |
| "loss": 0.4421, |
| "step": 225 |
| }, |
| { |
| "epoch": 3.4466666666666668, |
| "grad_norm": 0.1465432249850391, |
| "learning_rate": 2.0624346166633425e-05, |
| "loss": 0.4396, |
| "step": 226 |
| }, |
| { |
| "epoch": 3.461904761904762, |
| "grad_norm": 0.12349201931738121, |
| "learning_rate": 2.0248977993600912e-05, |
| "loss": 0.4491, |
| "step": 227 |
| }, |
| { |
| "epoch": 3.4771428571428573, |
| "grad_norm": 0.14981067132432477, |
| "learning_rate": 1.987589604797227e-05, |
| "loss": 0.4448, |
| "step": 228 |
| }, |
| { |
| "epoch": 3.4923809523809526, |
| "grad_norm": 0.1137656845152472, |
| "learning_rate": 1.9505143514863056e-05, |
| "loss": 0.4489, |
| "step": 229 |
| }, |
| { |
| "epoch": 3.5076190476190474, |
| "grad_norm": 0.1398849247762614, |
| "learning_rate": 1.913676330975385e-05, |
| "loss": 0.4546, |
| "step": 230 |
| }, |
| { |
| "epoch": 3.522857142857143, |
| "grad_norm": 0.12820525613752842, |
| "learning_rate": 1.8770798073522657e-05, |
| "loss": 0.447, |
| "step": 231 |
| }, |
| { |
| "epoch": 3.538095238095238, |
| "grad_norm": 0.13773852598915348, |
| "learning_rate": 1.8407290167509163e-05, |
| "loss": 0.4543, |
| "step": 232 |
| }, |
| { |
| "epoch": 3.5533333333333332, |
| "grad_norm": 0.12262425169332791, |
| "learning_rate": 1.8046281668611242e-05, |
| "loss": 0.444, |
| "step": 233 |
| }, |
| { |
| "epoch": 3.5685714285714285, |
| "grad_norm": 0.11207677036541486, |
| "learning_rate": 1.7687814364414508e-05, |
| "loss": 0.4451, |
| "step": 234 |
| }, |
| { |
| "epoch": 3.583809523809524, |
| "grad_norm": 0.11456531045490899, |
| "learning_rate": 1.73319297483553e-05, |
| "loss": 0.4465, |
| "step": 235 |
| }, |
| { |
| "epoch": 3.599047619047619, |
| "grad_norm": 0.12011483524657235, |
| "learning_rate": 1.6978669014917723e-05, |
| "loss": 0.4406, |
| "step": 236 |
| }, |
| { |
| "epoch": 3.6142857142857143, |
| "grad_norm": 0.11200693227331737, |
| "learning_rate": 1.6628073054865262e-05, |
| "loss": 0.4539, |
| "step": 237 |
| }, |
| { |
| "epoch": 3.6295238095238096, |
| "grad_norm": 0.11945653235133909, |
| "learning_rate": 1.6280182450507593e-05, |
| "loss": 0.4505, |
| "step": 238 |
| }, |
| { |
| "epoch": 3.644761904761905, |
| "grad_norm": 0.11328522908106976, |
| "learning_rate": 1.593503747100309e-05, |
| "loss": 0.44, |
| "step": 239 |
| }, |
| { |
| "epoch": 3.66, |
| "grad_norm": 0.11584353620347927, |
| "learning_rate": 1.5592678067697594e-05, |
| "loss": 0.4401, |
| "step": 240 |
| }, |
| { |
| "epoch": 3.675238095238095, |
| "grad_norm": 0.11338268232248719, |
| "learning_rate": 1.5253143869499875e-05, |
| "loss": 0.4526, |
| "step": 241 |
| }, |
| { |
| "epoch": 3.6904761904761907, |
| "grad_norm": 0.11825521687064344, |
| "learning_rate": 1.4916474178294578e-05, |
| "loss": 0.4472, |
| "step": 242 |
| }, |
| { |
| "epoch": 3.7057142857142855, |
| "grad_norm": 0.10393717835800384, |
| "learning_rate": 1.4582707964392909e-05, |
| "loss": 0.4503, |
| "step": 243 |
| }, |
| { |
| "epoch": 3.720952380952381, |
| "grad_norm": 0.11344323679399201, |
| "learning_rate": 1.4251883862021715e-05, |
| "loss": 0.4507, |
| "step": 244 |
| }, |
| { |
| "epoch": 3.736190476190476, |
| "grad_norm": 0.11795276330820621, |
| "learning_rate": 1.3924040164851497e-05, |
| "loss": 0.4509, |
| "step": 245 |
| }, |
| { |
| "epoch": 3.7514285714285713, |
| "grad_norm": 0.10446068928724489, |
| "learning_rate": 1.359921482156382e-05, |
| "loss": 0.453, |
| "step": 246 |
| }, |
| { |
| "epoch": 3.7666666666666666, |
| "grad_norm": 0.11963166659223025, |
| "learning_rate": 1.3277445431458653e-05, |
| "loss": 0.4423, |
| "step": 247 |
| }, |
| { |
| "epoch": 3.781904761904762, |
| "grad_norm": 0.11133235152901447, |
| "learning_rate": 1.2958769240102206e-05, |
| "loss": 0.4465, |
| "step": 248 |
| }, |
| { |
| "epoch": 3.797142857142857, |
| "grad_norm": 0.10488189725095455, |
| "learning_rate": 1.264322313501559e-05, |
| "loss": 0.4495, |
| "step": 249 |
| }, |
| { |
| "epoch": 3.8123809523809524, |
| "grad_norm": 0.09788048647298295, |
| "learning_rate": 1.2330843641405057e-05, |
| "loss": 0.4451, |
| "step": 250 |
| }, |
| { |
| "epoch": 3.8276190476190477, |
| "grad_norm": 0.1068932171231341, |
| "learning_rate": 1.2021666917934125e-05, |
| "loss": 0.4473, |
| "step": 251 |
| }, |
| { |
| "epoch": 3.842857142857143, |
| "grad_norm": 0.09965606097681808, |
| "learning_rate": 1.1715728752538103e-05, |
| "loss": 0.4443, |
| "step": 252 |
| }, |
| { |
| "epoch": 3.8580952380952382, |
| "grad_norm": 0.10053631876045145, |
| "learning_rate": 1.1413064558281564e-05, |
| "loss": 0.4468, |
| "step": 253 |
| }, |
| { |
| "epoch": 3.873333333333333, |
| "grad_norm": 0.09801079877642455, |
| "learning_rate": 1.1113709369259187e-05, |
| "loss": 0.4426, |
| "step": 254 |
| }, |
| { |
| "epoch": 3.888571428571429, |
| "grad_norm": 0.09483370655189176, |
| "learning_rate": 1.0817697836540493e-05, |
| "loss": 0.4493, |
| "step": 255 |
| }, |
| { |
| "epoch": 3.9038095238095236, |
| "grad_norm": 0.09358289511323944, |
| "learning_rate": 1.052506422415891e-05, |
| "loss": 0.4415, |
| "step": 256 |
| }, |
| { |
| "epoch": 3.919047619047619, |
| "grad_norm": 0.09937735008586661, |
| "learning_rate": 1.0235842405145587e-05, |
| "loss": 0.447, |
| "step": 257 |
| }, |
| { |
| "epoch": 3.934285714285714, |
| "grad_norm": 0.09713725112426527, |
| "learning_rate": 9.950065857608537e-06, |
| "loss": 0.4449, |
| "step": 258 |
| }, |
| { |
| "epoch": 3.9495238095238094, |
| "grad_norm": 0.10061824714900855, |
| "learning_rate": 9.667767660857508e-06, |
| "loss": 0.4378, |
| "step": 259 |
| }, |
| { |
| "epoch": 3.9647619047619047, |
| "grad_norm": 0.09975526240303212, |
| "learning_rate": 9.388980491574901e-06, |
| "loss": 0.4485, |
| "step": 260 |
| }, |
| { |
| "epoch": 3.98, |
| "grad_norm": 0.09583315993538538, |
| "learning_rate": 9.11373662003343e-06, |
| "loss": 0.4518, |
| "step": 261 |
| }, |
| { |
| "epoch": 3.9952380952380953, |
| "grad_norm": 0.12153372041343533, |
| "learning_rate": 8.842067906360716e-06, |
| "loss": 0.559, |
| "step": 262 |
| }, |
| { |
| "epoch": 4.011428571428572, |
| "grad_norm": 0.14086095004056465, |
| "learning_rate": 8.574005796851419e-06, |
| "loss": 0.5283, |
| "step": 263 |
| }, |
| { |
| "epoch": 4.026666666666666, |
| "grad_norm": 0.11788004209905123, |
| "learning_rate": 8.309581320327282e-06, |
| "loss": 0.4248, |
| "step": 264 |
| }, |
| { |
| "epoch": 4.041904761904762, |
| "grad_norm": 0.11207857563036874, |
| "learning_rate": 8.048825084545413e-06, |
| "loss": 0.4208, |
| "step": 265 |
| }, |
| { |
| "epoch": 4.057142857142857, |
| "grad_norm": 0.11130142780853505, |
| "learning_rate": 7.791767272655403e-06, |
| "loss": 0.4282, |
| "step": 266 |
| }, |
| { |
| "epoch": 4.072380952380953, |
| "grad_norm": 0.11636579346899405, |
| "learning_rate": 7.5384376397055644e-06, |
| "loss": 0.4275, |
| "step": 267 |
| }, |
| { |
| "epoch": 4.0876190476190475, |
| "grad_norm": 0.12911191155727517, |
| "learning_rate": 7.2888655091986995e-06, |
| "loss": 0.4296, |
| "step": 268 |
| }, |
| { |
| "epoch": 4.102857142857143, |
| "grad_norm": 0.1207757027022162, |
| "learning_rate": 7.043079769697833e-06, |
| "loss": 0.4301, |
| "step": 269 |
| }, |
| { |
| "epoch": 4.118095238095238, |
| "grad_norm": 0.10942499693335449, |
| "learning_rate": 6.801108871482287e-06, |
| "loss": 0.432, |
| "step": 270 |
| }, |
| { |
| "epoch": 4.133333333333334, |
| "grad_norm": 0.11446276362313004, |
| "learning_rate": 6.562980823254519e-06, |
| "loss": 0.4313, |
| "step": 271 |
| }, |
| { |
| "epoch": 4.148571428571429, |
| "grad_norm": 0.11355557887705738, |
| "learning_rate": 6.328723188898029e-06, |
| "loss": 0.4279, |
| "step": 272 |
| }, |
| { |
| "epoch": 4.163809523809523, |
| "grad_norm": 0.1062761145201427, |
| "learning_rate": 6.098363084286765e-06, |
| "loss": 0.4259, |
| "step": 273 |
| }, |
| { |
| "epoch": 4.179047619047619, |
| "grad_norm": 0.10273617911128365, |
| "learning_rate": 5.8719271741464056e-06, |
| "loss": 0.4225, |
| "step": 274 |
| }, |
| { |
| "epoch": 4.194285714285714, |
| "grad_norm": 0.10086062111767927, |
| "learning_rate": 5.6494416689678855e-06, |
| "loss": 0.4189, |
| "step": 275 |
| }, |
| { |
| "epoch": 4.20952380952381, |
| "grad_norm": 0.09559781846200713, |
| "learning_rate": 5.430932321973425e-06, |
| "loss": 0.4158, |
| "step": 276 |
| }, |
| { |
| "epoch": 4.2247619047619045, |
| "grad_norm": 0.09465874774217785, |
| "learning_rate": 5.21642442613556e-06, |
| "loss": 0.4299, |
| "step": 277 |
| }, |
| { |
| "epoch": 4.24, |
| "grad_norm": 0.09527880228230132, |
| "learning_rate": 5.005942811249412e-06, |
| "loss": 0.4243, |
| "step": 278 |
| }, |
| { |
| "epoch": 4.255238095238095, |
| "grad_norm": 0.09428002792877421, |
| "learning_rate": 4.799511841058575e-06, |
| "loss": 0.4262, |
| "step": 279 |
| }, |
| { |
| "epoch": 4.270476190476191, |
| "grad_norm": 0.09763553814742547, |
| "learning_rate": 4.597155410434972e-06, |
| "loss": 0.4247, |
| "step": 280 |
| }, |
| { |
| "epoch": 4.285714285714286, |
| "grad_norm": 0.09287137908820702, |
| "learning_rate": 4.398896942612934e-06, |
| "loss": 0.4238, |
| "step": 281 |
| }, |
| { |
| "epoch": 4.300952380952381, |
| "grad_norm": 0.09888369046987704, |
| "learning_rate": 4.204759386477961e-06, |
| "loss": 0.4391, |
| "step": 282 |
| }, |
| { |
| "epoch": 4.316190476190476, |
| "grad_norm": 0.09609824332962343, |
| "learning_rate": 4.01476521391027e-06, |
| "loss": 0.4352, |
| "step": 283 |
| }, |
| { |
| "epoch": 4.331428571428571, |
| "grad_norm": 0.09233308942677047, |
| "learning_rate": 3.828936417183671e-06, |
| "loss": 0.43, |
| "step": 284 |
| }, |
| { |
| "epoch": 4.346666666666667, |
| "grad_norm": 0.0890288964909307, |
| "learning_rate": 3.6472945064198827e-06, |
| "loss": 0.427, |
| "step": 285 |
| }, |
| { |
| "epoch": 4.3619047619047615, |
| "grad_norm": 0.08989519773450554, |
| "learning_rate": 3.4698605070987077e-06, |
| "loss": 0.4257, |
| "step": 286 |
| }, |
| { |
| "epoch": 4.377142857142857, |
| "grad_norm": 0.09526649707374434, |
| "learning_rate": 3.2966549576242477e-06, |
| "loss": 0.4357, |
| "step": 287 |
| }, |
| { |
| "epoch": 4.392380952380952, |
| "grad_norm": 0.09131870660723178, |
| "learning_rate": 3.127697906947584e-06, |
| "loss": 0.4272, |
| "step": 288 |
| }, |
| { |
| "epoch": 4.407619047619048, |
| "grad_norm": 0.08945756240639792, |
| "learning_rate": 2.963008912246008e-06, |
| "loss": 0.4347, |
| "step": 289 |
| }, |
| { |
| "epoch": 4.422857142857143, |
| "grad_norm": 0.08483664610068463, |
| "learning_rate": 2.802607036659284e-06, |
| "loss": 0.4294, |
| "step": 290 |
| }, |
| { |
| "epoch": 4.438095238095238, |
| "grad_norm": 0.08486549550090346, |
| "learning_rate": 2.646510847082997e-06, |
| "loss": 0.4303, |
| "step": 291 |
| }, |
| { |
| "epoch": 4.453333333333333, |
| "grad_norm": 0.084355078151828, |
| "learning_rate": 2.494738412019442e-06, |
| "loss": 0.423, |
| "step": 292 |
| }, |
| { |
| "epoch": 4.468571428571429, |
| "grad_norm": 0.08518364314817348, |
| "learning_rate": 2.3473072994861167e-06, |
| "loss": 0.427, |
| "step": 293 |
| }, |
| { |
| "epoch": 4.483809523809524, |
| "grad_norm": 0.08287931866258626, |
| "learning_rate": 2.2042345749821913e-06, |
| "loss": 0.4216, |
| "step": 294 |
| }, |
| { |
| "epoch": 4.499047619047619, |
| "grad_norm": 0.08289838691379334, |
| "learning_rate": 2.065536799513135e-06, |
| "loss": 0.4288, |
| "step": 295 |
| }, |
| { |
| "epoch": 4.514285714285714, |
| "grad_norm": 0.0850326818939402, |
| "learning_rate": 1.9312300276737516e-06, |
| "loss": 0.4235, |
| "step": 296 |
| }, |
| { |
| "epoch": 4.529523809523809, |
| "grad_norm": 0.08392478938826013, |
| "learning_rate": 1.8013298057897932e-06, |
| "loss": 0.4293, |
| "step": 297 |
| }, |
| { |
| "epoch": 4.544761904761905, |
| "grad_norm": 0.0892781985318678, |
| "learning_rate": 1.6758511701184809e-06, |
| "loss": 0.4238, |
| "step": 298 |
| }, |
| { |
| "epoch": 4.5600000000000005, |
| "grad_norm": 0.08180406525124756, |
| "learning_rate": 1.5548086451079747e-06, |
| "loss": 0.4384, |
| "step": 299 |
| }, |
| { |
| "epoch": 4.575238095238095, |
| "grad_norm": 0.08011550631601072, |
| "learning_rate": 1.4382162417161748e-06, |
| "loss": 0.4206, |
| "step": 300 |
| }, |
| { |
| "epoch": 4.59047619047619, |
| "grad_norm": 0.08066581660175723, |
| "learning_rate": 1.3260874557889002e-06, |
| "loss": 0.4287, |
| "step": 301 |
| }, |
| { |
| "epoch": 4.605714285714286, |
| "grad_norm": 0.0806215224586864, |
| "learning_rate": 1.2184352664977105e-06, |
| "loss": 0.4175, |
| "step": 302 |
| }, |
| { |
| "epoch": 4.620952380952381, |
| "grad_norm": 0.07969848747023248, |
| "learning_rate": 1.115272134837544e-06, |
| "loss": 0.4312, |
| "step": 303 |
| }, |
| { |
| "epoch": 4.636190476190476, |
| "grad_norm": 0.0797095157185979, |
| "learning_rate": 1.0166100021843283e-06, |
| "loss": 0.4346, |
| "step": 304 |
| }, |
| { |
| "epoch": 4.651428571428571, |
| "grad_norm": 0.0792961159548818, |
| "learning_rate": 9.224602889127321e-07, |
| "loss": 0.4286, |
| "step": 305 |
| }, |
| { |
| "epoch": 4.666666666666667, |
| "grad_norm": 0.08285323866652071, |
| "learning_rate": 8.328338930742296e-07, |
| "loss": 0.4267, |
| "step": 306 |
| }, |
| { |
| "epoch": 4.681904761904762, |
| "grad_norm": 0.07823875843722773, |
| "learning_rate": 7.477411891356268e-07, |
| "loss": 0.4276, |
| "step": 307 |
| }, |
| { |
| "epoch": 4.6971428571428575, |
| "grad_norm": 0.07763022247982207, |
| "learning_rate": 6.671920267782029e-07, |
| "loss": 0.4282, |
| "step": 308 |
| }, |
| { |
| "epoch": 4.712380952380952, |
| "grad_norm": 0.07871343041787499, |
| "learning_rate": 5.911957297575743e-07, |
| "loss": 0.4323, |
| "step": 309 |
| }, |
| { |
| "epoch": 4.727619047619047, |
| "grad_norm": 0.07681136995775628, |
| "learning_rate": 5.197610948244469e-07, |
| "loss": 0.4256, |
| "step": 310 |
| }, |
| { |
| "epoch": 4.742857142857143, |
| "grad_norm": 0.07665250706053243, |
| "learning_rate": 4.5289639070638103e-07, |
| "loss": 0.4171, |
| "step": 311 |
| }, |
| { |
| "epoch": 4.758095238095238, |
| "grad_norm": 0.07835687105047397, |
| "learning_rate": 3.906093571506597e-07, |
| "loss": 0.428, |
| "step": 312 |
| }, |
| { |
| "epoch": 4.773333333333333, |
| "grad_norm": 0.07949715755664716, |
| "learning_rate": 3.329072040284009e-07, |
| "loss": 0.4208, |
| "step": 313 |
| }, |
| { |
| "epoch": 4.788571428571428, |
| "grad_norm": 0.0773084553588578, |
| "learning_rate": 2.797966104999805e-07, |
| "loss": 0.4332, |
| "step": 314 |
| }, |
| { |
| "epoch": 4.803809523809524, |
| "grad_norm": 0.07795013678746698, |
| "learning_rate": 2.3128372424192635e-07, |
| "loss": 0.4377, |
| "step": 315 |
| }, |
| { |
| "epoch": 4.819047619047619, |
| "grad_norm": 0.07666311315306534, |
| "learning_rate": 1.8737416073529635e-07, |
| "loss": 0.4272, |
| "step": 316 |
| }, |
| { |
| "epoch": 4.8342857142857145, |
| "grad_norm": 0.07693103942282913, |
| "learning_rate": 1.4807300261568735e-07, |
| "loss": 0.4287, |
| "step": 317 |
| }, |
| { |
| "epoch": 4.849523809523809, |
| "grad_norm": 0.08092920067828374, |
| "learning_rate": 1.1338479908488798e-07, |
| "loss": 0.4337, |
| "step": 318 |
| }, |
| { |
| "epoch": 4.864761904761905, |
| "grad_norm": 0.07921080037622698, |
| "learning_rate": 8.331356538431313e-08, |
| "loss": 0.4262, |
| "step": 319 |
| }, |
| { |
| "epoch": 4.88, |
| "grad_norm": 0.08216408996820164, |
| "learning_rate": 5.7862782330215804e-08, |
| "loss": 0.4317, |
| "step": 320 |
| }, |
| { |
| "epoch": 4.895238095238096, |
| "grad_norm": 0.07591222440722502, |
| "learning_rate": 3.703539591080052e-08, |
| "loss": 0.4232, |
| "step": 321 |
| }, |
| { |
| "epoch": 4.91047619047619, |
| "grad_norm": 0.08045135907636206, |
| "learning_rate": 2.0833816945184936e-08, |
| "loss": 0.4297, |
| "step": 322 |
| }, |
| { |
| "epoch": 4.925714285714285, |
| "grad_norm": 0.07684321088822434, |
| "learning_rate": 9.25992080436533e-09, |
| "loss": 0.4276, |
| "step": 323 |
| }, |
| { |
| "epoch": 4.940952380952381, |
| "grad_norm": 0.07680665420588856, |
| "learning_rate": 2.315047194136888e-09, |
| "loss": 0.4271, |
| "step": 324 |
| }, |
| { |
| "epoch": 4.956190476190476, |
| "grad_norm": 0.07663586105720761, |
| "learning_rate": 0.0, |
| "loss": 0.4248, |
| "step": 325 |
| }, |
| { |
| "epoch": 4.956190476190476, |
| "step": 325, |
| "total_flos": 8.641809758367515e+18, |
| "train_loss": 0.5190348960803105, |
| "train_runtime": 75842.3225, |
| "train_samples_per_second": 2.214, |
| "train_steps_per_second": 0.004 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 325, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 8.641809758367515e+18, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|