{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.1765536723163842, "eval_steps": 500, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00035310734463276836, "grad_norm": 0.45211324095726013, "learning_rate": 5e-05, "loss": 3.273, "step": 1 }, { "epoch": 0.0007062146892655367, "grad_norm": 0.30982279777526855, "learning_rate": 4.99e-05, "loss": 2.5125, "step": 2 }, { "epoch": 0.001059322033898305, "grad_norm": 0.39342790842056274, "learning_rate": 4.9800000000000004e-05, "loss": 2.8347, "step": 3 }, { "epoch": 0.0014124293785310734, "grad_norm": 0.3912785053253174, "learning_rate": 4.97e-05, "loss": 2.6189, "step": 4 }, { "epoch": 0.0017655367231638418, "grad_norm": 0.43560200929641724, "learning_rate": 4.96e-05, "loss": 2.8784, "step": 5 }, { "epoch": 0.00211864406779661, "grad_norm": 0.5163903832435608, "learning_rate": 4.9500000000000004e-05, "loss": 3.0496, "step": 6 }, { "epoch": 0.0024717514124293787, "grad_norm": 0.3364083468914032, "learning_rate": 4.94e-05, "loss": 2.0403, "step": 7 }, { "epoch": 0.002824858757062147, "grad_norm": 0.4723369777202606, "learning_rate": 4.93e-05, "loss": 2.5065, "step": 8 }, { "epoch": 0.0031779661016949155, "grad_norm": 0.5686460137367249, "learning_rate": 4.92e-05, "loss": 2.8529, "step": 9 }, { "epoch": 0.0035310734463276836, "grad_norm": 0.7666249871253967, "learning_rate": 4.91e-05, "loss": 3.2163, "step": 10 }, { "epoch": 0.003884180790960452, "grad_norm": 0.5175825953483582, "learning_rate": 4.9e-05, "loss": 2.3817, "step": 11 }, { "epoch": 0.00423728813559322, "grad_norm": 0.7693667411804199, "learning_rate": 4.89e-05, "loss": 3.1173, "step": 12 }, { "epoch": 0.004590395480225989, "grad_norm": 0.6010498404502869, "learning_rate": 4.88e-05, "loss": 2.5264, "step": 13 }, { "epoch": 0.0049435028248587575, "grad_norm": 0.8724003434181213, "learning_rate": 4.87e-05, "loss": 3.1545, "step": 14 }, { "epoch": 0.005296610169491525, "grad_norm": 0.6222332119941711, "learning_rate": 4.86e-05, "loss": 2.4859, "step": 15 }, { "epoch": 0.005649717514124294, "grad_norm": 0.5855898261070251, "learning_rate": 4.85e-05, "loss": 2.2748, "step": 16 }, { "epoch": 0.006002824858757062, "grad_norm": 0.8227508068084717, "learning_rate": 4.8400000000000004e-05, "loss": 2.777, "step": 17 }, { "epoch": 0.006355932203389831, "grad_norm": 0.747832179069519, "learning_rate": 4.83e-05, "loss": 2.6429, "step": 18 }, { "epoch": 0.006709039548022599, "grad_norm": 0.8245562314987183, "learning_rate": 4.82e-05, "loss": 2.8865, "step": 19 }, { "epoch": 0.007062146892655367, "grad_norm": 0.8959988951683044, "learning_rate": 4.8100000000000004e-05, "loss": 2.8727, "step": 20 }, { "epoch": 0.007415254237288136, "grad_norm": 0.7936757802963257, "learning_rate": 4.8e-05, "loss": 2.5376, "step": 21 }, { "epoch": 0.007768361581920904, "grad_norm": 0.8037199378013611, "learning_rate": 4.79e-05, "loss": 2.6522, "step": 22 }, { "epoch": 0.008121468926553672, "grad_norm": 0.7104085087776184, "learning_rate": 4.78e-05, "loss": 2.3427, "step": 23 }, { "epoch": 0.00847457627118644, "grad_norm": 0.8215932846069336, "learning_rate": 4.77e-05, "loss": 2.7918, "step": 24 }, { "epoch": 0.00882768361581921, "grad_norm": 1.0282623767852783, "learning_rate": 4.76e-05, "loss": 2.9389, "step": 25 }, { "epoch": 0.009180790960451978, "grad_norm": 0.7932501435279846, "learning_rate": 4.75e-05, "loss": 2.7691, "step": 26 }, { "epoch": 0.009533898305084746, "grad_norm": 0.6620106101036072, "learning_rate": 4.74e-05, "loss": 2.1335, "step": 27 }, { "epoch": 0.009887005649717515, "grad_norm": 0.8058229684829712, "learning_rate": 4.73e-05, "loss": 2.4778, "step": 28 }, { "epoch": 0.010240112994350282, "grad_norm": 0.7049869894981384, "learning_rate": 4.72e-05, "loss": 2.3379, "step": 29 }, { "epoch": 0.01059322033898305, "grad_norm": 0.6863004565238953, "learning_rate": 4.71e-05, "loss": 2.0131, "step": 30 }, { "epoch": 0.010946327683615819, "grad_norm": 0.9368977546691895, "learning_rate": 4.7e-05, "loss": 2.4664, "step": 31 }, { "epoch": 0.011299435028248588, "grad_norm": 0.7291913628578186, "learning_rate": 4.69e-05, "loss": 2.2541, "step": 32 }, { "epoch": 0.011652542372881356, "grad_norm": 0.6416012048721313, "learning_rate": 4.6800000000000006e-05, "loss": 1.8727, "step": 33 }, { "epoch": 0.012005649717514125, "grad_norm": 0.6020755767822266, "learning_rate": 4.6700000000000003e-05, "loss": 1.8596, "step": 34 }, { "epoch": 0.012358757062146893, "grad_norm": 0.635847270488739, "learning_rate": 4.660000000000001e-05, "loss": 1.9688, "step": 35 }, { "epoch": 0.012711864406779662, "grad_norm": 0.6946876645088196, "learning_rate": 4.6500000000000005e-05, "loss": 2.6249, "step": 36 }, { "epoch": 0.013064971751412429, "grad_norm": 0.7043133974075317, "learning_rate": 4.64e-05, "loss": 2.5754, "step": 37 }, { "epoch": 0.013418079096045197, "grad_norm": 0.5472213625907898, "learning_rate": 4.630000000000001e-05, "loss": 1.91, "step": 38 }, { "epoch": 0.013771186440677966, "grad_norm": 0.5201235413551331, "learning_rate": 4.6200000000000005e-05, "loss": 2.0296, "step": 39 }, { "epoch": 0.014124293785310734, "grad_norm": 0.5103974342346191, "learning_rate": 4.61e-05, "loss": 1.7567, "step": 40 }, { "epoch": 0.014477401129943503, "grad_norm": 0.42137110233306885, "learning_rate": 4.600000000000001e-05, "loss": 1.7719, "step": 41 }, { "epoch": 0.014830508474576272, "grad_norm": 0.6730819940567017, "learning_rate": 4.5900000000000004e-05, "loss": 2.4245, "step": 42 }, { "epoch": 0.01518361581920904, "grad_norm": 0.6370269060134888, "learning_rate": 4.58e-05, "loss": 2.501, "step": 43 }, { "epoch": 0.015536723163841809, "grad_norm": 0.5084403157234192, "learning_rate": 4.5700000000000006e-05, "loss": 2.0893, "step": 44 }, { "epoch": 0.015889830508474576, "grad_norm": 0.6537813544273376, "learning_rate": 4.5600000000000004e-05, "loss": 2.3681, "step": 45 }, { "epoch": 0.016242937853107344, "grad_norm": 0.5805037617683411, "learning_rate": 4.55e-05, "loss": 2.3028, "step": 46 }, { "epoch": 0.016596045197740113, "grad_norm": 0.5214577913284302, "learning_rate": 4.5400000000000006e-05, "loss": 2.1601, "step": 47 }, { "epoch": 0.01694915254237288, "grad_norm": 0.7660169005393982, "learning_rate": 4.53e-05, "loss": 2.4195, "step": 48 }, { "epoch": 0.01730225988700565, "grad_norm": 0.537522554397583, "learning_rate": 4.52e-05, "loss": 2.3828, "step": 49 }, { "epoch": 0.01765536723163842, "grad_norm": 0.5875993371009827, "learning_rate": 4.5100000000000005e-05, "loss": 2.4529, "step": 50 }, { "epoch": 0.018008474576271187, "grad_norm": 0.480882465839386, "learning_rate": 4.5e-05, "loss": 1.7377, "step": 51 }, { "epoch": 0.018361581920903956, "grad_norm": 0.4598454236984253, "learning_rate": 4.49e-05, "loss": 1.8277, "step": 52 }, { "epoch": 0.018714689265536724, "grad_norm": 0.4913174510002136, "learning_rate": 4.4800000000000005e-05, "loss": 2.0882, "step": 53 }, { "epoch": 0.019067796610169493, "grad_norm": 0.4395110011100769, "learning_rate": 4.47e-05, "loss": 1.8825, "step": 54 }, { "epoch": 0.01942090395480226, "grad_norm": 0.5200910568237305, "learning_rate": 4.46e-05, "loss": 2.1883, "step": 55 }, { "epoch": 0.01977401129943503, "grad_norm": 0.7119576930999756, "learning_rate": 4.4500000000000004e-05, "loss": 2.3147, "step": 56 }, { "epoch": 0.020127118644067795, "grad_norm": 0.5865232348442078, "learning_rate": 4.44e-05, "loss": 2.4823, "step": 57 }, { "epoch": 0.020480225988700564, "grad_norm": 0.5342106223106384, "learning_rate": 4.43e-05, "loss": 2.3112, "step": 58 }, { "epoch": 0.020833333333333332, "grad_norm": 0.63221275806427, "learning_rate": 4.4200000000000004e-05, "loss": 2.4642, "step": 59 }, { "epoch": 0.0211864406779661, "grad_norm": 0.5381503701210022, "learning_rate": 4.41e-05, "loss": 2.1702, "step": 60 }, { "epoch": 0.02153954802259887, "grad_norm": 0.34933093190193176, "learning_rate": 4.4000000000000006e-05, "loss": 1.6217, "step": 61 }, { "epoch": 0.021892655367231638, "grad_norm": 0.4486427307128906, "learning_rate": 4.39e-05, "loss": 1.9545, "step": 62 }, { "epoch": 0.022245762711864406, "grad_norm": 0.39275842905044556, "learning_rate": 4.38e-05, "loss": 1.692, "step": 63 }, { "epoch": 0.022598870056497175, "grad_norm": 0.5706101655960083, "learning_rate": 4.3700000000000005e-05, "loss": 2.3182, "step": 64 }, { "epoch": 0.022951977401129944, "grad_norm": 0.43103381991386414, "learning_rate": 4.36e-05, "loss": 1.8674, "step": 65 }, { "epoch": 0.023305084745762712, "grad_norm": 0.6326637268066406, "learning_rate": 4.35e-05, "loss": 2.3281, "step": 66 }, { "epoch": 0.02365819209039548, "grad_norm": 0.5602039694786072, "learning_rate": 4.3400000000000005e-05, "loss": 2.24, "step": 67 }, { "epoch": 0.02401129943502825, "grad_norm": 0.42276912927627563, "learning_rate": 4.33e-05, "loss": 1.9933, "step": 68 }, { "epoch": 0.024364406779661018, "grad_norm": 0.5240011811256409, "learning_rate": 4.32e-05, "loss": 2.1384, "step": 69 }, { "epoch": 0.024717514124293787, "grad_norm": 0.579456627368927, "learning_rate": 4.3100000000000004e-05, "loss": 2.2305, "step": 70 }, { "epoch": 0.025070621468926555, "grad_norm": 0.5782936215400696, "learning_rate": 4.3e-05, "loss": 2.2021, "step": 71 }, { "epoch": 0.025423728813559324, "grad_norm": 0.6299965977668762, "learning_rate": 4.29e-05, "loss": 2.3214, "step": 72 }, { "epoch": 0.02577683615819209, "grad_norm": 0.634605884552002, "learning_rate": 4.2800000000000004e-05, "loss": 2.3198, "step": 73 }, { "epoch": 0.026129943502824857, "grad_norm": 0.5833197832107544, "learning_rate": 4.27e-05, "loss": 2.1705, "step": 74 }, { "epoch": 0.026483050847457626, "grad_norm": 0.5400241017341614, "learning_rate": 4.26e-05, "loss": 1.8534, "step": 75 }, { "epoch": 0.026836158192090395, "grad_norm": 0.6428772807121277, "learning_rate": 4.25e-05, "loss": 2.3044, "step": 76 }, { "epoch": 0.027189265536723163, "grad_norm": 0.611422061920166, "learning_rate": 4.24e-05, "loss": 2.1591, "step": 77 }, { "epoch": 0.02754237288135593, "grad_norm": 0.602570652961731, "learning_rate": 4.23e-05, "loss": 2.1249, "step": 78 }, { "epoch": 0.0278954802259887, "grad_norm": 0.48029282689094543, "learning_rate": 4.22e-05, "loss": 1.905, "step": 79 }, { "epoch": 0.02824858757062147, "grad_norm": 0.6572628617286682, "learning_rate": 4.21e-05, "loss": 2.2358, "step": 80 }, { "epoch": 0.028601694915254237, "grad_norm": 0.5220134258270264, "learning_rate": 4.2e-05, "loss": 1.9954, "step": 81 }, { "epoch": 0.028954802259887006, "grad_norm": 0.4906415641307831, "learning_rate": 4.19e-05, "loss": 1.8642, "step": 82 }, { "epoch": 0.029307909604519775, "grad_norm": 0.6776078343391418, "learning_rate": 4.18e-05, "loss": 2.2061, "step": 83 }, { "epoch": 0.029661016949152543, "grad_norm": 0.496427059173584, "learning_rate": 4.17e-05, "loss": 1.9494, "step": 84 }, { "epoch": 0.03001412429378531, "grad_norm": 0.4432898759841919, "learning_rate": 4.16e-05, "loss": 1.6809, "step": 85 }, { "epoch": 0.03036723163841808, "grad_norm": 0.6649871468544006, "learning_rate": 4.15e-05, "loss": 2.0615, "step": 86 }, { "epoch": 0.03072033898305085, "grad_norm": 0.6379185318946838, "learning_rate": 4.14e-05, "loss": 1.9796, "step": 87 }, { "epoch": 0.031073446327683617, "grad_norm": 0.5551559925079346, "learning_rate": 4.13e-05, "loss": 1.874, "step": 88 }, { "epoch": 0.031426553672316386, "grad_norm": 0.7269681096076965, "learning_rate": 4.12e-05, "loss": 2.1407, "step": 89 }, { "epoch": 0.03177966101694915, "grad_norm": 0.6741663217544556, "learning_rate": 4.11e-05, "loss": 1.9975, "step": 90 }, { "epoch": 0.03213276836158192, "grad_norm": 0.5726102590560913, "learning_rate": 4.1e-05, "loss": 1.8897, "step": 91 }, { "epoch": 0.03248587570621469, "grad_norm": 0.7033076286315918, "learning_rate": 4.09e-05, "loss": 2.0044, "step": 92 }, { "epoch": 0.03283898305084746, "grad_norm": 0.6314510107040405, "learning_rate": 4.08e-05, "loss": 2.1114, "step": 93 }, { "epoch": 0.033192090395480225, "grad_norm": 0.49348896741867065, "learning_rate": 4.07e-05, "loss": 1.5142, "step": 94 }, { "epoch": 0.033545197740113, "grad_norm": 0.5636611580848694, "learning_rate": 4.0600000000000004e-05, "loss": 1.7442, "step": 95 }, { "epoch": 0.03389830508474576, "grad_norm": 0.5502078533172607, "learning_rate": 4.05e-05, "loss": 1.744, "step": 96 }, { "epoch": 0.03425141242937853, "grad_norm": 0.7853062152862549, "learning_rate": 4.0400000000000006e-05, "loss": 2.0471, "step": 97 }, { "epoch": 0.0346045197740113, "grad_norm": 0.7374521493911743, "learning_rate": 4.0300000000000004e-05, "loss": 1.9658, "step": 98 }, { "epoch": 0.034957627118644065, "grad_norm": 0.5843427181243896, "learning_rate": 4.02e-05, "loss": 1.698, "step": 99 }, { "epoch": 0.03531073446327684, "grad_norm": 0.7103281617164612, "learning_rate": 4.0100000000000006e-05, "loss": 1.8531, "step": 100 }, { "epoch": 0.0356638418079096, "grad_norm": 0.547626256942749, "learning_rate": 4e-05, "loss": 1.6113, "step": 101 }, { "epoch": 0.036016949152542374, "grad_norm": 0.7051151990890503, "learning_rate": 3.99e-05, "loss": 1.8774, "step": 102 }, { "epoch": 0.03637005649717514, "grad_norm": 0.5148686766624451, "learning_rate": 3.9800000000000005e-05, "loss": 1.6082, "step": 103 }, { "epoch": 0.03672316384180791, "grad_norm": 0.5013210773468018, "learning_rate": 3.97e-05, "loss": 1.6964, "step": 104 }, { "epoch": 0.037076271186440676, "grad_norm": 0.7677639126777649, "learning_rate": 3.960000000000001e-05, "loss": 1.8114, "step": 105 }, { "epoch": 0.03742937853107345, "grad_norm": 0.6067560911178589, "learning_rate": 3.9500000000000005e-05, "loss": 1.6947, "step": 106 }, { "epoch": 0.037782485875706213, "grad_norm": 0.5241538286209106, "learning_rate": 3.94e-05, "loss": 1.4953, "step": 107 }, { "epoch": 0.038135593220338986, "grad_norm": 0.5037181377410889, "learning_rate": 3.9300000000000007e-05, "loss": 1.4166, "step": 108 }, { "epoch": 0.03848870056497175, "grad_norm": 0.7840654253959656, "learning_rate": 3.9200000000000004e-05, "loss": 1.8501, "step": 109 }, { "epoch": 0.03884180790960452, "grad_norm": 0.6080065369606018, "learning_rate": 3.91e-05, "loss": 1.6206, "step": 110 }, { "epoch": 0.03919491525423729, "grad_norm": 0.616134762763977, "learning_rate": 3.9000000000000006e-05, "loss": 1.7034, "step": 111 }, { "epoch": 0.03954802259887006, "grad_norm": 0.7034599781036377, "learning_rate": 3.8900000000000004e-05, "loss": 1.9318, "step": 112 }, { "epoch": 0.039901129943502825, "grad_norm": 0.8047641515731812, "learning_rate": 3.88e-05, "loss": 1.7351, "step": 113 }, { "epoch": 0.04025423728813559, "grad_norm": 0.8917104005813599, "learning_rate": 3.8700000000000006e-05, "loss": 1.8398, "step": 114 }, { "epoch": 0.04060734463276836, "grad_norm": 0.8026954531669617, "learning_rate": 3.86e-05, "loss": 1.7842, "step": 115 }, { "epoch": 0.04096045197740113, "grad_norm": 0.5217316150665283, "learning_rate": 3.85e-05, "loss": 1.5885, "step": 116 }, { "epoch": 0.0413135593220339, "grad_norm": 0.6288496851921082, "learning_rate": 3.8400000000000005e-05, "loss": 1.5425, "step": 117 }, { "epoch": 0.041666666666666664, "grad_norm": 0.77554851770401, "learning_rate": 3.83e-05, "loss": 1.6508, "step": 118 }, { "epoch": 0.042019774011299436, "grad_norm": 0.5694034695625305, "learning_rate": 3.82e-05, "loss": 1.2894, "step": 119 }, { "epoch": 0.0423728813559322, "grad_norm": 0.6931347846984863, "learning_rate": 3.8100000000000005e-05, "loss": 1.7874, "step": 120 }, { "epoch": 0.042725988700564974, "grad_norm": 0.7180691957473755, "learning_rate": 3.8e-05, "loss": 1.5679, "step": 121 }, { "epoch": 0.04307909604519774, "grad_norm": 0.5494984984397888, "learning_rate": 3.79e-05, "loss": 1.2885, "step": 122 }, { "epoch": 0.04343220338983051, "grad_norm": 0.8779198527336121, "learning_rate": 3.7800000000000004e-05, "loss": 1.6815, "step": 123 }, { "epoch": 0.043785310734463276, "grad_norm": 0.7112749218940735, "learning_rate": 3.77e-05, "loss": 1.4889, "step": 124 }, { "epoch": 0.04413841807909605, "grad_norm": 0.6239631772041321, "learning_rate": 3.76e-05, "loss": 1.3156, "step": 125 }, { "epoch": 0.04449152542372881, "grad_norm": 0.8934807181358337, "learning_rate": 3.7500000000000003e-05, "loss": 1.7629, "step": 126 }, { "epoch": 0.044844632768361585, "grad_norm": 0.9005821943283081, "learning_rate": 3.74e-05, "loss": 1.5775, "step": 127 }, { "epoch": 0.04519774011299435, "grad_norm": 0.7520875334739685, "learning_rate": 3.73e-05, "loss": 1.6556, "step": 128 }, { "epoch": 0.045550847457627115, "grad_norm": 0.8343580961227417, "learning_rate": 3.72e-05, "loss": 1.5429, "step": 129 }, { "epoch": 0.04590395480225989, "grad_norm": 0.7819499373435974, "learning_rate": 3.71e-05, "loss": 1.5313, "step": 130 }, { "epoch": 0.04625706214689265, "grad_norm": 0.6105003952980042, "learning_rate": 3.7e-05, "loss": 1.3124, "step": 131 }, { "epoch": 0.046610169491525424, "grad_norm": 0.6769589185714722, "learning_rate": 3.69e-05, "loss": 1.4305, "step": 132 }, { "epoch": 0.04696327683615819, "grad_norm": 0.7654910683631897, "learning_rate": 3.68e-05, "loss": 1.4972, "step": 133 }, { "epoch": 0.04731638418079096, "grad_norm": 0.7610757946968079, "learning_rate": 3.6700000000000004e-05, "loss": 1.4652, "step": 134 }, { "epoch": 0.04766949152542373, "grad_norm": 0.6594064831733704, "learning_rate": 3.66e-05, "loss": 1.2301, "step": 135 }, { "epoch": 0.0480225988700565, "grad_norm": 0.719152569770813, "learning_rate": 3.65e-05, "loss": 1.3399, "step": 136 }, { "epoch": 0.048375706214689264, "grad_norm": 0.9311054348945618, "learning_rate": 3.6400000000000004e-05, "loss": 1.5786, "step": 137 }, { "epoch": 0.048728813559322036, "grad_norm": 0.7845869660377502, "learning_rate": 3.63e-05, "loss": 1.3293, "step": 138 }, { "epoch": 0.0490819209039548, "grad_norm": 0.6555350422859192, "learning_rate": 3.62e-05, "loss": 1.1677, "step": 139 }, { "epoch": 0.04943502824858757, "grad_norm": 0.863808810710907, "learning_rate": 3.61e-05, "loss": 1.3453, "step": 140 }, { "epoch": 0.04978813559322034, "grad_norm": 1.1489522457122803, "learning_rate": 3.6e-05, "loss": 1.5277, "step": 141 }, { "epoch": 0.05014124293785311, "grad_norm": 0.8987497091293335, "learning_rate": 3.59e-05, "loss": 1.3435, "step": 142 }, { "epoch": 0.050494350282485875, "grad_norm": 1.1033629179000854, "learning_rate": 3.58e-05, "loss": 1.5565, "step": 143 }, { "epoch": 0.05084745762711865, "grad_norm": 0.9892997145652771, "learning_rate": 3.57e-05, "loss": 1.5017, "step": 144 }, { "epoch": 0.05120056497175141, "grad_norm": 0.8673040866851807, "learning_rate": 3.56e-05, "loss": 1.3484, "step": 145 }, { "epoch": 0.05155367231638418, "grad_norm": 0.8514487147331238, "learning_rate": 3.55e-05, "loss": 1.4329, "step": 146 }, { "epoch": 0.05190677966101695, "grad_norm": 0.7865382432937622, "learning_rate": 3.54e-05, "loss": 1.3205, "step": 147 }, { "epoch": 0.052259887005649715, "grad_norm": 0.9893730878829956, "learning_rate": 3.53e-05, "loss": 1.2557, "step": 148 }, { "epoch": 0.05261299435028249, "grad_norm": 0.7459471225738525, "learning_rate": 3.52e-05, "loss": 1.398, "step": 149 }, { "epoch": 0.05296610169491525, "grad_norm": 1.2132583856582642, "learning_rate": 3.51e-05, "loss": 1.396, "step": 150 }, { "epoch": 0.053319209039548024, "grad_norm": 1.1160212755203247, "learning_rate": 3.5e-05, "loss": 1.3013, "step": 151 }, { "epoch": 0.05367231638418079, "grad_norm": 1.058889389038086, "learning_rate": 3.49e-05, "loss": 1.4149, "step": 152 }, { "epoch": 0.05402542372881356, "grad_norm": 1.065194845199585, "learning_rate": 3.48e-05, "loss": 1.3913, "step": 153 }, { "epoch": 0.054378531073446326, "grad_norm": 0.8159513473510742, "learning_rate": 3.4699999999999996e-05, "loss": 1.1171, "step": 154 }, { "epoch": 0.0547316384180791, "grad_norm": 1.2298649549484253, "learning_rate": 3.46e-05, "loss": 1.2559, "step": 155 }, { "epoch": 0.05508474576271186, "grad_norm": 0.845851480960846, "learning_rate": 3.45e-05, "loss": 1.1559, "step": 156 }, { "epoch": 0.055437853107344635, "grad_norm": 0.8105614185333252, "learning_rate": 3.4399999999999996e-05, "loss": 1.092, "step": 157 }, { "epoch": 0.0557909604519774, "grad_norm": 0.9624556303024292, "learning_rate": 3.430000000000001e-05, "loss": 1.1862, "step": 158 }, { "epoch": 0.05614406779661017, "grad_norm": 1.2695869207382202, "learning_rate": 3.4200000000000005e-05, "loss": 1.1886, "step": 159 }, { "epoch": 0.05649717514124294, "grad_norm": 0.9481382369995117, "learning_rate": 3.41e-05, "loss": 1.2366, "step": 160 }, { "epoch": 0.0568502824858757, "grad_norm": 1.1612368822097778, "learning_rate": 3.4000000000000007e-05, "loss": 1.3102, "step": 161 }, { "epoch": 0.057203389830508475, "grad_norm": 1.296866536140442, "learning_rate": 3.3900000000000004e-05, "loss": 1.2689, "step": 162 }, { "epoch": 0.05755649717514124, "grad_norm": 1.0605868101119995, "learning_rate": 3.38e-05, "loss": 1.1056, "step": 163 }, { "epoch": 0.05790960451977401, "grad_norm": 1.06467604637146, "learning_rate": 3.3700000000000006e-05, "loss": 1.2329, "step": 164 }, { "epoch": 0.05826271186440678, "grad_norm": 0.9635260105133057, "learning_rate": 3.3600000000000004e-05, "loss": 1.082, "step": 165 }, { "epoch": 0.05861581920903955, "grad_norm": 1.2294679880142212, "learning_rate": 3.35e-05, "loss": 1.3892, "step": 166 }, { "epoch": 0.058968926553672314, "grad_norm": 1.4516520500183105, "learning_rate": 3.3400000000000005e-05, "loss": 1.0772, "step": 167 }, { "epoch": 0.059322033898305086, "grad_norm": 1.08195960521698, "learning_rate": 3.33e-05, "loss": 1.0811, "step": 168 }, { "epoch": 0.05967514124293785, "grad_norm": 1.1554380655288696, "learning_rate": 3.32e-05, "loss": 1.1417, "step": 169 }, { "epoch": 0.06002824858757062, "grad_norm": 1.117101788520813, "learning_rate": 3.3100000000000005e-05, "loss": 1.2163, "step": 170 }, { "epoch": 0.06038135593220339, "grad_norm": 1.2141193151474, "learning_rate": 3.3e-05, "loss": 1.2469, "step": 171 }, { "epoch": 0.06073446327683616, "grad_norm": 1.29922616481781, "learning_rate": 3.29e-05, "loss": 1.2541, "step": 172 }, { "epoch": 0.061087570621468926, "grad_norm": 1.3265364170074463, "learning_rate": 3.2800000000000004e-05, "loss": 1.1884, "step": 173 }, { "epoch": 0.0614406779661017, "grad_norm": 1.4586164951324463, "learning_rate": 3.27e-05, "loss": 1.1861, "step": 174 }, { "epoch": 0.06179378531073446, "grad_norm": 1.3976298570632935, "learning_rate": 3.26e-05, "loss": 1.0568, "step": 175 }, { "epoch": 0.062146892655367235, "grad_norm": 1.0947189331054688, "learning_rate": 3.2500000000000004e-05, "loss": 1.0583, "step": 176 }, { "epoch": 0.0625, "grad_norm": 1.0739431381225586, "learning_rate": 3.24e-05, "loss": 0.973, "step": 177 }, { "epoch": 0.06285310734463277, "grad_norm": 1.0500564575195312, "learning_rate": 3.2300000000000006e-05, "loss": 0.961, "step": 178 }, { "epoch": 0.06320621468926553, "grad_norm": 1.3906699419021606, "learning_rate": 3.2200000000000003e-05, "loss": 1.1366, "step": 179 }, { "epoch": 0.0635593220338983, "grad_norm": 1.2436065673828125, "learning_rate": 3.21e-05, "loss": 1.0769, "step": 180 }, { "epoch": 0.06391242937853107, "grad_norm": 1.137102484703064, "learning_rate": 3.2000000000000005e-05, "loss": 0.9609, "step": 181 }, { "epoch": 0.06426553672316385, "grad_norm": 1.7101986408233643, "learning_rate": 3.19e-05, "loss": 1.138, "step": 182 }, { "epoch": 0.0646186440677966, "grad_norm": 1.1683790683746338, "learning_rate": 3.18e-05, "loss": 1.1872, "step": 183 }, { "epoch": 0.06497175141242938, "grad_norm": 1.6510800123214722, "learning_rate": 3.1700000000000005e-05, "loss": 1.0618, "step": 184 }, { "epoch": 0.06532485875706215, "grad_norm": 1.5926707983016968, "learning_rate": 3.16e-05, "loss": 1.0558, "step": 185 }, { "epoch": 0.06567796610169492, "grad_norm": 1.502793312072754, "learning_rate": 3.15e-05, "loss": 0.9155, "step": 186 }, { "epoch": 0.06603107344632768, "grad_norm": 1.5513490438461304, "learning_rate": 3.1400000000000004e-05, "loss": 1.1119, "step": 187 }, { "epoch": 0.06638418079096045, "grad_norm": 1.7766902446746826, "learning_rate": 3.13e-05, "loss": 1.0724, "step": 188 }, { "epoch": 0.06673728813559322, "grad_norm": 1.6195155382156372, "learning_rate": 3.12e-05, "loss": 1.0377, "step": 189 }, { "epoch": 0.067090395480226, "grad_norm": 1.5613833665847778, "learning_rate": 3.1100000000000004e-05, "loss": 1.0057, "step": 190 }, { "epoch": 0.06744350282485875, "grad_norm": 1.5648159980773926, "learning_rate": 3.1e-05, "loss": 0.9758, "step": 191 }, { "epoch": 0.06779661016949153, "grad_norm": 2.3179798126220703, "learning_rate": 3.09e-05, "loss": 1.076, "step": 192 }, { "epoch": 0.0681497175141243, "grad_norm": 1.6775708198547363, "learning_rate": 3.08e-05, "loss": 0.8133, "step": 193 }, { "epoch": 0.06850282485875706, "grad_norm": 1.6693114042282104, "learning_rate": 3.07e-05, "loss": 0.8958, "step": 194 }, { "epoch": 0.06885593220338983, "grad_norm": 1.623583436012268, "learning_rate": 3.06e-05, "loss": 0.9226, "step": 195 }, { "epoch": 0.0692090395480226, "grad_norm": 1.6135281324386597, "learning_rate": 3.05e-05, "loss": 0.8707, "step": 196 }, { "epoch": 0.06956214689265537, "grad_norm": 1.6629116535186768, "learning_rate": 3.04e-05, "loss": 0.8883, "step": 197 }, { "epoch": 0.06991525423728813, "grad_norm": 1.9516435861587524, "learning_rate": 3.03e-05, "loss": 0.8909, "step": 198 }, { "epoch": 0.0702683615819209, "grad_norm": 1.9378143548965454, "learning_rate": 3.02e-05, "loss": 0.7697, "step": 199 }, { "epoch": 0.07062146892655367, "grad_norm": 2.0332353115081787, "learning_rate": 3.01e-05, "loss": 0.9999, "step": 200 }, { "epoch": 0.07097457627118645, "grad_norm": 1.4929075241088867, "learning_rate": 3e-05, "loss": 0.8301, "step": 201 }, { "epoch": 0.0713276836158192, "grad_norm": 1.9575985670089722, "learning_rate": 2.9900000000000002e-05, "loss": 1.0613, "step": 202 }, { "epoch": 0.07168079096045198, "grad_norm": 1.3801037073135376, "learning_rate": 2.98e-05, "loss": 1.0363, "step": 203 }, { "epoch": 0.07203389830508475, "grad_norm": 1.8629417419433594, "learning_rate": 2.97e-05, "loss": 0.8268, "step": 204 }, { "epoch": 0.07238700564971752, "grad_norm": 1.5077078342437744, "learning_rate": 2.96e-05, "loss": 0.9769, "step": 205 }, { "epoch": 0.07274011299435028, "grad_norm": 1.8674854040145874, "learning_rate": 2.95e-05, "loss": 1.011, "step": 206 }, { "epoch": 0.07309322033898305, "grad_norm": 2.605397939682007, "learning_rate": 2.94e-05, "loss": 1.2138, "step": 207 }, { "epoch": 0.07344632768361582, "grad_norm": 1.3890961408615112, "learning_rate": 2.93e-05, "loss": 0.9088, "step": 208 }, { "epoch": 0.0737994350282486, "grad_norm": 2.1312317848205566, "learning_rate": 2.9199999999999998e-05, "loss": 1.2366, "step": 209 }, { "epoch": 0.07415254237288135, "grad_norm": 1.817219853401184, "learning_rate": 2.91e-05, "loss": 0.7755, "step": 210 }, { "epoch": 0.07450564971751412, "grad_norm": 1.8591907024383545, "learning_rate": 2.9e-05, "loss": 0.9953, "step": 211 }, { "epoch": 0.0748587570621469, "grad_norm": 1.7456353902816772, "learning_rate": 2.8899999999999998e-05, "loss": 0.7605, "step": 212 }, { "epoch": 0.07521186440677965, "grad_norm": 1.6678094863891602, "learning_rate": 2.88e-05, "loss": 0.779, "step": 213 }, { "epoch": 0.07556497175141243, "grad_norm": 1.7547659873962402, "learning_rate": 2.87e-05, "loss": 0.7277, "step": 214 }, { "epoch": 0.0759180790960452, "grad_norm": 2.3935739994049072, "learning_rate": 2.86e-05, "loss": 1.0243, "step": 215 }, { "epoch": 0.07627118644067797, "grad_norm": 2.1737802028656006, "learning_rate": 2.8499999999999998e-05, "loss": 0.8159, "step": 216 }, { "epoch": 0.07662429378531073, "grad_norm": 2.551464557647705, "learning_rate": 2.84e-05, "loss": 0.7808, "step": 217 }, { "epoch": 0.0769774011299435, "grad_norm": 2.0867929458618164, "learning_rate": 2.83e-05, "loss": 0.7764, "step": 218 }, { "epoch": 0.07733050847457627, "grad_norm": 2.001920700073242, "learning_rate": 2.8199999999999998e-05, "loss": 0.7499, "step": 219 }, { "epoch": 0.07768361581920905, "grad_norm": 2.4141225814819336, "learning_rate": 2.8100000000000005e-05, "loss": 0.7964, "step": 220 }, { "epoch": 0.0780367231638418, "grad_norm": 2.371626615524292, "learning_rate": 2.8000000000000003e-05, "loss": 0.8997, "step": 221 }, { "epoch": 0.07838983050847458, "grad_norm": 1.6987398862838745, "learning_rate": 2.7900000000000004e-05, "loss": 0.9769, "step": 222 }, { "epoch": 0.07874293785310735, "grad_norm": 1.768283486366272, "learning_rate": 2.7800000000000005e-05, "loss": 0.6651, "step": 223 }, { "epoch": 0.07909604519774012, "grad_norm": 1.827958106994629, "learning_rate": 2.7700000000000002e-05, "loss": 0.6805, "step": 224 }, { "epoch": 0.07944915254237288, "grad_norm": 2.7550013065338135, "learning_rate": 2.7600000000000003e-05, "loss": 0.8034, "step": 225 }, { "epoch": 0.07980225988700565, "grad_norm": 1.7423927783966064, "learning_rate": 2.7500000000000004e-05, "loss": 0.7345, "step": 226 }, { "epoch": 0.08015536723163842, "grad_norm": 1.9607951641082764, "learning_rate": 2.7400000000000002e-05, "loss": 0.6771, "step": 227 }, { "epoch": 0.08050847457627118, "grad_norm": 2.737659215927124, "learning_rate": 2.7300000000000003e-05, "loss": 1.1056, "step": 228 }, { "epoch": 0.08086158192090395, "grad_norm": 3.3492250442504883, "learning_rate": 2.7200000000000004e-05, "loss": 1.2373, "step": 229 }, { "epoch": 0.08121468926553672, "grad_norm": 2.10353946685791, "learning_rate": 2.7100000000000005e-05, "loss": 0.8509, "step": 230 }, { "epoch": 0.0815677966101695, "grad_norm": 2.1462435722351074, "learning_rate": 2.7000000000000002e-05, "loss": 0.8893, "step": 231 }, { "epoch": 0.08192090395480225, "grad_norm": 2.2392938137054443, "learning_rate": 2.6900000000000003e-05, "loss": 0.8682, "step": 232 }, { "epoch": 0.08227401129943503, "grad_norm": 2.242868661880493, "learning_rate": 2.6800000000000004e-05, "loss": 0.812, "step": 233 }, { "epoch": 0.0826271186440678, "grad_norm": 1.9220484495162964, "learning_rate": 2.6700000000000002e-05, "loss": 0.7693, "step": 234 }, { "epoch": 0.08298022598870057, "grad_norm": 1.5621178150177002, "learning_rate": 2.6600000000000003e-05, "loss": 0.875, "step": 235 }, { "epoch": 0.08333333333333333, "grad_norm": 1.8653405904769897, "learning_rate": 2.6500000000000004e-05, "loss": 0.6442, "step": 236 }, { "epoch": 0.0836864406779661, "grad_norm": 2.035147190093994, "learning_rate": 2.64e-05, "loss": 0.8103, "step": 237 }, { "epoch": 0.08403954802259887, "grad_norm": 2.0776431560516357, "learning_rate": 2.6300000000000002e-05, "loss": 0.8276, "step": 238 }, { "epoch": 0.08439265536723164, "grad_norm": 1.745665192604065, "learning_rate": 2.6200000000000003e-05, "loss": 0.6145, "step": 239 }, { "epoch": 0.0847457627118644, "grad_norm": 1.9675729274749756, "learning_rate": 2.61e-05, "loss": 0.7005, "step": 240 }, { "epoch": 0.08509887005649718, "grad_norm": 2.2235677242279053, "learning_rate": 2.6000000000000002e-05, "loss": 0.5884, "step": 241 }, { "epoch": 0.08545197740112995, "grad_norm": 1.8204203844070435, "learning_rate": 2.5900000000000003e-05, "loss": 0.6029, "step": 242 }, { "epoch": 0.0858050847457627, "grad_norm": 1.6106022596359253, "learning_rate": 2.58e-05, "loss": 0.5933, "step": 243 }, { "epoch": 0.08615819209039548, "grad_norm": 1.8598597049713135, "learning_rate": 2.57e-05, "loss": 0.6792, "step": 244 }, { "epoch": 0.08651129943502825, "grad_norm": 2.186007499694824, "learning_rate": 2.5600000000000002e-05, "loss": 0.8079, "step": 245 }, { "epoch": 0.08686440677966102, "grad_norm": 2.206789255142212, "learning_rate": 2.5500000000000003e-05, "loss": 0.936, "step": 246 }, { "epoch": 0.08721751412429378, "grad_norm": 1.865030288696289, "learning_rate": 2.54e-05, "loss": 0.8269, "step": 247 }, { "epoch": 0.08757062146892655, "grad_norm": 3.2279410362243652, "learning_rate": 2.5300000000000002e-05, "loss": 0.9072, "step": 248 }, { "epoch": 0.08792372881355932, "grad_norm": 1.5364686250686646, "learning_rate": 2.5200000000000003e-05, "loss": 0.5486, "step": 249 }, { "epoch": 0.0882768361581921, "grad_norm": 1.7777973413467407, "learning_rate": 2.51e-05, "loss": 0.5444, "step": 250 }, { "epoch": 0.08862994350282485, "grad_norm": 2.1404268741607666, "learning_rate": 2.5e-05, "loss": 0.6368, "step": 251 }, { "epoch": 0.08898305084745763, "grad_norm": 1.80747389793396, "learning_rate": 2.4900000000000002e-05, "loss": 0.6483, "step": 252 }, { "epoch": 0.0893361581920904, "grad_norm": 1.780600666999817, "learning_rate": 2.48e-05, "loss": 0.7834, "step": 253 }, { "epoch": 0.08968926553672317, "grad_norm": 2.232827663421631, "learning_rate": 2.47e-05, "loss": 0.7617, "step": 254 }, { "epoch": 0.09004237288135593, "grad_norm": 1.9321049451828003, "learning_rate": 2.46e-05, "loss": 0.5936, "step": 255 }, { "epoch": 0.0903954802259887, "grad_norm": 1.7767024040222168, "learning_rate": 2.45e-05, "loss": 0.5475, "step": 256 }, { "epoch": 0.09074858757062147, "grad_norm": 1.7460103034973145, "learning_rate": 2.44e-05, "loss": 0.7268, "step": 257 }, { "epoch": 0.09110169491525423, "grad_norm": 2.010267734527588, "learning_rate": 2.43e-05, "loss": 0.5198, "step": 258 }, { "epoch": 0.091454802259887, "grad_norm": 2.2814278602600098, "learning_rate": 2.4200000000000002e-05, "loss": 0.8076, "step": 259 }, { "epoch": 0.09180790960451977, "grad_norm": 2.043018341064453, "learning_rate": 2.41e-05, "loss": 0.7212, "step": 260 }, { "epoch": 0.09216101694915255, "grad_norm": 1.5442239046096802, "learning_rate": 2.4e-05, "loss": 0.6238, "step": 261 }, { "epoch": 0.0925141242937853, "grad_norm": 2.0493950843811035, "learning_rate": 2.39e-05, "loss": 0.6279, "step": 262 }, { "epoch": 0.09286723163841808, "grad_norm": 1.9058549404144287, "learning_rate": 2.38e-05, "loss": 0.6636, "step": 263 }, { "epoch": 0.09322033898305085, "grad_norm": 1.766516089439392, "learning_rate": 2.37e-05, "loss": 0.7014, "step": 264 }, { "epoch": 0.09357344632768362, "grad_norm": 1.7538968324661255, "learning_rate": 2.36e-05, "loss": 0.7425, "step": 265 }, { "epoch": 0.09392655367231638, "grad_norm": 1.9766415357589722, "learning_rate": 2.35e-05, "loss": 0.6857, "step": 266 }, { "epoch": 0.09427966101694915, "grad_norm": 1.8969347476959229, "learning_rate": 2.3400000000000003e-05, "loss": 0.509, "step": 267 }, { "epoch": 0.09463276836158192, "grad_norm": 2.556140422821045, "learning_rate": 2.3300000000000004e-05, "loss": 0.9246, "step": 268 }, { "epoch": 0.0949858757062147, "grad_norm": 2.192314624786377, "learning_rate": 2.32e-05, "loss": 0.8475, "step": 269 }, { "epoch": 0.09533898305084745, "grad_norm": 1.6340404748916626, "learning_rate": 2.3100000000000002e-05, "loss": 0.8045, "step": 270 }, { "epoch": 0.09569209039548023, "grad_norm": 1.849310040473938, "learning_rate": 2.3000000000000003e-05, "loss": 0.6461, "step": 271 }, { "epoch": 0.096045197740113, "grad_norm": 1.7643516063690186, "learning_rate": 2.29e-05, "loss": 0.8039, "step": 272 }, { "epoch": 0.09639830508474577, "grad_norm": 2.249030351638794, "learning_rate": 2.2800000000000002e-05, "loss": 0.6257, "step": 273 }, { "epoch": 0.09675141242937853, "grad_norm": 1.8955878019332886, "learning_rate": 2.2700000000000003e-05, "loss": 0.4904, "step": 274 }, { "epoch": 0.0971045197740113, "grad_norm": 2.037266254425049, "learning_rate": 2.26e-05, "loss": 0.9272, "step": 275 }, { "epoch": 0.09745762711864407, "grad_norm": 2.061370611190796, "learning_rate": 2.25e-05, "loss": 0.5984, "step": 276 }, { "epoch": 0.09781073446327683, "grad_norm": 3.4211597442626953, "learning_rate": 2.2400000000000002e-05, "loss": 0.9706, "step": 277 }, { "epoch": 0.0981638418079096, "grad_norm": 2.7527401447296143, "learning_rate": 2.23e-05, "loss": 0.7066, "step": 278 }, { "epoch": 0.09851694915254237, "grad_norm": 1.8547669649124146, "learning_rate": 2.22e-05, "loss": 0.458, "step": 279 }, { "epoch": 0.09887005649717515, "grad_norm": 1.9559056758880615, "learning_rate": 2.2100000000000002e-05, "loss": 0.453, "step": 280 }, { "epoch": 0.0992231638418079, "grad_norm": 1.7714954614639282, "learning_rate": 2.2000000000000003e-05, "loss": 0.5063, "step": 281 }, { "epoch": 0.09957627118644068, "grad_norm": 3.084165334701538, "learning_rate": 2.19e-05, "loss": 0.8611, "step": 282 }, { "epoch": 0.09992937853107345, "grad_norm": 1.8296337127685547, "learning_rate": 2.18e-05, "loss": 0.6884, "step": 283 }, { "epoch": 0.10028248587570622, "grad_norm": 2.5133588314056396, "learning_rate": 2.1700000000000002e-05, "loss": 0.6772, "step": 284 }, { "epoch": 0.10063559322033898, "grad_norm": 1.612042784690857, "learning_rate": 2.16e-05, "loss": 0.4726, "step": 285 }, { "epoch": 0.10098870056497175, "grad_norm": 1.9312748908996582, "learning_rate": 2.15e-05, "loss": 0.6412, "step": 286 }, { "epoch": 0.10134180790960452, "grad_norm": 2.4446702003479004, "learning_rate": 2.1400000000000002e-05, "loss": 0.8463, "step": 287 }, { "epoch": 0.1016949152542373, "grad_norm": 4.253321647644043, "learning_rate": 2.13e-05, "loss": 0.6177, "step": 288 }, { "epoch": 0.10204802259887005, "grad_norm": 2.584301471710205, "learning_rate": 2.12e-05, "loss": 0.7182, "step": 289 }, { "epoch": 0.10240112994350282, "grad_norm": 1.7066371440887451, "learning_rate": 2.11e-05, "loss": 0.5563, "step": 290 }, { "epoch": 0.1027542372881356, "grad_norm": 3.1515045166015625, "learning_rate": 2.1e-05, "loss": 0.8663, "step": 291 }, { "epoch": 0.10310734463276836, "grad_norm": 2.3074467182159424, "learning_rate": 2.09e-05, "loss": 0.8209, "step": 292 }, { "epoch": 0.10346045197740113, "grad_norm": 2.0748472213745117, "learning_rate": 2.08e-05, "loss": 0.6029, "step": 293 }, { "epoch": 0.1038135593220339, "grad_norm": 1.8115911483764648, "learning_rate": 2.07e-05, "loss": 0.4101, "step": 294 }, { "epoch": 0.10416666666666667, "grad_norm": 4.644301891326904, "learning_rate": 2.06e-05, "loss": 0.8823, "step": 295 }, { "epoch": 0.10451977401129943, "grad_norm": 1.8356761932373047, "learning_rate": 2.05e-05, "loss": 0.6229, "step": 296 }, { "epoch": 0.1048728813559322, "grad_norm": 2.584825038909912, "learning_rate": 2.04e-05, "loss": 0.7403, "step": 297 }, { "epoch": 0.10522598870056497, "grad_norm": 2.1190598011016846, "learning_rate": 2.0300000000000002e-05, "loss": 0.5843, "step": 298 }, { "epoch": 0.10557909604519775, "grad_norm": 2.0052149295806885, "learning_rate": 2.0200000000000003e-05, "loss": 0.5476, "step": 299 }, { "epoch": 0.1059322033898305, "grad_norm": 2.5021302700042725, "learning_rate": 2.01e-05, "loss": 0.6322, "step": 300 }, { "epoch": 0.10628531073446328, "grad_norm": 2.790252208709717, "learning_rate": 2e-05, "loss": 0.7356, "step": 301 }, { "epoch": 0.10663841807909605, "grad_norm": 1.896161437034607, "learning_rate": 1.9900000000000003e-05, "loss": 0.6414, "step": 302 }, { "epoch": 0.10699152542372882, "grad_norm": 2.121238946914673, "learning_rate": 1.9800000000000004e-05, "loss": 0.4369, "step": 303 }, { "epoch": 0.10734463276836158, "grad_norm": 2.465498685836792, "learning_rate": 1.97e-05, "loss": 0.7176, "step": 304 }, { "epoch": 0.10769774011299435, "grad_norm": 2.068796396255493, "learning_rate": 1.9600000000000002e-05, "loss": 0.4301, "step": 305 }, { "epoch": 0.10805084745762712, "grad_norm": 1.9807084798812866, "learning_rate": 1.9500000000000003e-05, "loss": 0.4117, "step": 306 }, { "epoch": 0.10840395480225988, "grad_norm": 2.1290664672851562, "learning_rate": 1.94e-05, "loss": 0.636, "step": 307 }, { "epoch": 0.10875706214689265, "grad_norm": 1.9659547805786133, "learning_rate": 1.93e-05, "loss": 0.51, "step": 308 }, { "epoch": 0.10911016949152542, "grad_norm": 2.1764960289001465, "learning_rate": 1.9200000000000003e-05, "loss": 0.6201, "step": 309 }, { "epoch": 0.1094632768361582, "grad_norm": 2.654648542404175, "learning_rate": 1.91e-05, "loss": 0.7892, "step": 310 }, { "epoch": 0.10981638418079095, "grad_norm": 1.8828105926513672, "learning_rate": 1.9e-05, "loss": 0.6453, "step": 311 }, { "epoch": 0.11016949152542373, "grad_norm": 2.0999748706817627, "learning_rate": 1.8900000000000002e-05, "loss": 0.7547, "step": 312 }, { "epoch": 0.1105225988700565, "grad_norm": 1.9529651403427124, "learning_rate": 1.88e-05, "loss": 0.3852, "step": 313 }, { "epoch": 0.11087570621468927, "grad_norm": 2.1099841594696045, "learning_rate": 1.87e-05, "loss": 0.8429, "step": 314 }, { "epoch": 0.11122881355932203, "grad_norm": 2.6168792247772217, "learning_rate": 1.86e-05, "loss": 0.8384, "step": 315 }, { "epoch": 0.1115819209039548, "grad_norm": 1.8378658294677734, "learning_rate": 1.85e-05, "loss": 0.5949, "step": 316 }, { "epoch": 0.11193502824858757, "grad_norm": 2.333801746368408, "learning_rate": 1.84e-05, "loss": 0.4418, "step": 317 }, { "epoch": 0.11228813559322035, "grad_norm": 2.6359400749206543, "learning_rate": 1.83e-05, "loss": 0.7196, "step": 318 }, { "epoch": 0.1126412429378531, "grad_norm": 1.830614447593689, "learning_rate": 1.8200000000000002e-05, "loss": 0.5038, "step": 319 }, { "epoch": 0.11299435028248588, "grad_norm": 1.8708122968673706, "learning_rate": 1.81e-05, "loss": 0.356, "step": 320 }, { "epoch": 0.11334745762711865, "grad_norm": 1.8595423698425293, "learning_rate": 1.8e-05, "loss": 0.5855, "step": 321 }, { "epoch": 0.1137005649717514, "grad_norm": 2.0622501373291016, "learning_rate": 1.79e-05, "loss": 0.4762, "step": 322 }, { "epoch": 0.11405367231638418, "grad_norm": 1.8576161861419678, "learning_rate": 1.78e-05, "loss": 0.6245, "step": 323 }, { "epoch": 0.11440677966101695, "grad_norm": 2.014206886291504, "learning_rate": 1.77e-05, "loss": 0.4627, "step": 324 }, { "epoch": 0.11475988700564972, "grad_norm": 1.8058968782424927, "learning_rate": 1.76e-05, "loss": 0.3285, "step": 325 }, { "epoch": 0.11511299435028248, "grad_norm": 1.7399612665176392, "learning_rate": 1.75e-05, "loss": 0.4845, "step": 326 }, { "epoch": 0.11546610169491525, "grad_norm": 2.8591692447662354, "learning_rate": 1.74e-05, "loss": 0.7819, "step": 327 }, { "epoch": 0.11581920903954802, "grad_norm": 2.9054338932037354, "learning_rate": 1.73e-05, "loss": 0.6961, "step": 328 }, { "epoch": 0.1161723163841808, "grad_norm": 2.231865167617798, "learning_rate": 1.7199999999999998e-05, "loss": 0.532, "step": 329 }, { "epoch": 0.11652542372881355, "grad_norm": 3.112537145614624, "learning_rate": 1.7100000000000002e-05, "loss": 0.7317, "step": 330 }, { "epoch": 0.11687853107344633, "grad_norm": 1.886364221572876, "learning_rate": 1.7000000000000003e-05, "loss": 0.4278, "step": 331 }, { "epoch": 0.1172316384180791, "grad_norm": 2.2548704147338867, "learning_rate": 1.69e-05, "loss": 0.4959, "step": 332 }, { "epoch": 0.11758474576271187, "grad_norm": 1.778956651687622, "learning_rate": 1.6800000000000002e-05, "loss": 0.4093, "step": 333 }, { "epoch": 0.11793785310734463, "grad_norm": 3.506840229034424, "learning_rate": 1.6700000000000003e-05, "loss": 0.889, "step": 334 }, { "epoch": 0.1182909604519774, "grad_norm": 1.7313032150268555, "learning_rate": 1.66e-05, "loss": 0.3059, "step": 335 }, { "epoch": 0.11864406779661017, "grad_norm": 2.914469003677368, "learning_rate": 1.65e-05, "loss": 0.6975, "step": 336 }, { "epoch": 0.11899717514124294, "grad_norm": 2.3646645545959473, "learning_rate": 1.6400000000000002e-05, "loss": 0.3799, "step": 337 }, { "epoch": 0.1193502824858757, "grad_norm": 2.708645820617676, "learning_rate": 1.63e-05, "loss": 0.6651, "step": 338 }, { "epoch": 0.11970338983050847, "grad_norm": 2.4956207275390625, "learning_rate": 1.62e-05, "loss": 0.5777, "step": 339 }, { "epoch": 0.12005649717514125, "grad_norm": 2.371713161468506, "learning_rate": 1.6100000000000002e-05, "loss": 0.477, "step": 340 }, { "epoch": 0.120409604519774, "grad_norm": 1.81998872756958, "learning_rate": 1.6000000000000003e-05, "loss": 0.3214, "step": 341 }, { "epoch": 0.12076271186440678, "grad_norm": 2.2463088035583496, "learning_rate": 1.59e-05, "loss": 0.602, "step": 342 }, { "epoch": 0.12111581920903955, "grad_norm": 2.2001595497131348, "learning_rate": 1.58e-05, "loss": 0.6024, "step": 343 }, { "epoch": 0.12146892655367232, "grad_norm": 2.6866791248321533, "learning_rate": 1.5700000000000002e-05, "loss": 0.693, "step": 344 }, { "epoch": 0.12182203389830508, "grad_norm": 2.6667373180389404, "learning_rate": 1.56e-05, "loss": 0.6857, "step": 345 }, { "epoch": 0.12217514124293785, "grad_norm": 2.1822969913482666, "learning_rate": 1.55e-05, "loss": 0.51, "step": 346 }, { "epoch": 0.12252824858757062, "grad_norm": 2.881263256072998, "learning_rate": 1.54e-05, "loss": 0.6404, "step": 347 }, { "epoch": 0.1228813559322034, "grad_norm": 2.1057329177856445, "learning_rate": 1.53e-05, "loss": 0.495, "step": 348 }, { "epoch": 0.12323446327683615, "grad_norm": 1.6811401844024658, "learning_rate": 1.52e-05, "loss": 0.2848, "step": 349 }, { "epoch": 0.12358757062146893, "grad_norm": 2.3314602375030518, "learning_rate": 1.51e-05, "loss": 0.5686, "step": 350 }, { "epoch": 0.1239406779661017, "grad_norm": 2.8521921634674072, "learning_rate": 1.5e-05, "loss": 0.7763, "step": 351 }, { "epoch": 0.12429378531073447, "grad_norm": 2.379612922668457, "learning_rate": 1.49e-05, "loss": 0.6625, "step": 352 }, { "epoch": 0.12464689265536723, "grad_norm": 1.7267616987228394, "learning_rate": 1.48e-05, "loss": 0.3089, "step": 353 }, { "epoch": 0.125, "grad_norm": 2.078075647354126, "learning_rate": 1.47e-05, "loss": 0.5671, "step": 354 }, { "epoch": 0.12535310734463276, "grad_norm": 2.491132974624634, "learning_rate": 1.4599999999999999e-05, "loss": 0.5108, "step": 355 }, { "epoch": 0.12570621468926554, "grad_norm": 2.2118701934814453, "learning_rate": 1.45e-05, "loss": 0.4687, "step": 356 }, { "epoch": 0.1260593220338983, "grad_norm": 1.556470274925232, "learning_rate": 1.44e-05, "loss": 0.2878, "step": 357 }, { "epoch": 0.12641242937853106, "grad_norm": 3.0352628231048584, "learning_rate": 1.43e-05, "loss": 0.7796, "step": 358 }, { "epoch": 0.12676553672316385, "grad_norm": 2.498322010040283, "learning_rate": 1.42e-05, "loss": 0.6919, "step": 359 }, { "epoch": 0.1271186440677966, "grad_norm": 2.7435500621795654, "learning_rate": 1.4099999999999999e-05, "loss": 0.6379, "step": 360 }, { "epoch": 0.1274717514124294, "grad_norm": 2.1981303691864014, "learning_rate": 1.4000000000000001e-05, "loss": 0.4841, "step": 361 }, { "epoch": 0.12782485875706215, "grad_norm": 4.253507137298584, "learning_rate": 1.3900000000000002e-05, "loss": 0.8417, "step": 362 }, { "epoch": 0.1281779661016949, "grad_norm": 2.2498624324798584, "learning_rate": 1.3800000000000002e-05, "loss": 0.3849, "step": 363 }, { "epoch": 0.1285310734463277, "grad_norm": 2.5313210487365723, "learning_rate": 1.3700000000000001e-05, "loss": 0.7152, "step": 364 }, { "epoch": 0.12888418079096045, "grad_norm": 2.427729368209839, "learning_rate": 1.3600000000000002e-05, "loss": 0.613, "step": 365 }, { "epoch": 0.1292372881355932, "grad_norm": 2.1720805168151855, "learning_rate": 1.3500000000000001e-05, "loss": 0.527, "step": 366 }, { "epoch": 0.129590395480226, "grad_norm": 2.054292678833008, "learning_rate": 1.3400000000000002e-05, "loss": 0.4659, "step": 367 }, { "epoch": 0.12994350282485875, "grad_norm": 2.235578775405884, "learning_rate": 1.3300000000000001e-05, "loss": 0.4747, "step": 368 }, { "epoch": 0.13029661016949154, "grad_norm": 1.5781618356704712, "learning_rate": 1.32e-05, "loss": 0.2759, "step": 369 }, { "epoch": 0.1306497175141243, "grad_norm": 1.8499678373336792, "learning_rate": 1.3100000000000002e-05, "loss": 0.4529, "step": 370 }, { "epoch": 0.13100282485875706, "grad_norm": 1.901394248008728, "learning_rate": 1.3000000000000001e-05, "loss": 0.5179, "step": 371 }, { "epoch": 0.13135593220338984, "grad_norm": 2.1701114177703857, "learning_rate": 1.29e-05, "loss": 0.5104, "step": 372 }, { "epoch": 0.1317090395480226, "grad_norm": 2.1720824241638184, "learning_rate": 1.2800000000000001e-05, "loss": 0.3809, "step": 373 }, { "epoch": 0.13206214689265536, "grad_norm": 2.1540091037750244, "learning_rate": 1.27e-05, "loss": 0.5311, "step": 374 }, { "epoch": 0.13241525423728814, "grad_norm": 2.0049026012420654, "learning_rate": 1.2600000000000001e-05, "loss": 0.363, "step": 375 }, { "epoch": 0.1327683615819209, "grad_norm": 2.1072335243225098, "learning_rate": 1.25e-05, "loss": 0.4218, "step": 376 }, { "epoch": 0.13312146892655366, "grad_norm": 1.9690706729888916, "learning_rate": 1.24e-05, "loss": 0.4304, "step": 377 }, { "epoch": 0.13347457627118645, "grad_norm": 1.7669113874435425, "learning_rate": 1.23e-05, "loss": 0.3686, "step": 378 }, { "epoch": 0.1338276836158192, "grad_norm": 1.9775711297988892, "learning_rate": 1.22e-05, "loss": 0.503, "step": 379 }, { "epoch": 0.134180790960452, "grad_norm": 2.458343744277954, "learning_rate": 1.2100000000000001e-05, "loss": 0.4038, "step": 380 }, { "epoch": 0.13453389830508475, "grad_norm": 1.9653310775756836, "learning_rate": 1.2e-05, "loss": 0.5042, "step": 381 }, { "epoch": 0.1348870056497175, "grad_norm": 1.832046627998352, "learning_rate": 1.19e-05, "loss": 0.3696, "step": 382 }, { "epoch": 0.1352401129943503, "grad_norm": 4.0539374351501465, "learning_rate": 1.18e-05, "loss": 0.8044, "step": 383 }, { "epoch": 0.13559322033898305, "grad_norm": 4.322567462921143, "learning_rate": 1.1700000000000001e-05, "loss": 0.9251, "step": 384 }, { "epoch": 0.1359463276836158, "grad_norm": 1.9108883142471313, "learning_rate": 1.16e-05, "loss": 0.4827, "step": 385 }, { "epoch": 0.1362994350282486, "grad_norm": 1.4768140316009521, "learning_rate": 1.1500000000000002e-05, "loss": 0.2541, "step": 386 }, { "epoch": 0.13665254237288135, "grad_norm": 2.005392551422119, "learning_rate": 1.1400000000000001e-05, "loss": 0.5762, "step": 387 }, { "epoch": 0.1370056497175141, "grad_norm": 3.0048270225524902, "learning_rate": 1.13e-05, "loss": 0.6806, "step": 388 }, { "epoch": 0.1373587570621469, "grad_norm": 2.6098363399505615, "learning_rate": 1.1200000000000001e-05, "loss": 0.5409, "step": 389 }, { "epoch": 0.13771186440677965, "grad_norm": 2.0291426181793213, "learning_rate": 1.11e-05, "loss": 0.4466, "step": 390 }, { "epoch": 0.13806497175141244, "grad_norm": 1.4195945262908936, "learning_rate": 1.1000000000000001e-05, "loss": 0.2425, "step": 391 }, { "epoch": 0.1384180790960452, "grad_norm": 3.1550965309143066, "learning_rate": 1.09e-05, "loss": 0.6254, "step": 392 }, { "epoch": 0.13877118644067796, "grad_norm": 1.7930920124053955, "learning_rate": 1.08e-05, "loss": 0.3466, "step": 393 }, { "epoch": 0.13912429378531074, "grad_norm": 1.9969501495361328, "learning_rate": 1.0700000000000001e-05, "loss": 0.3996, "step": 394 }, { "epoch": 0.1394774011299435, "grad_norm": 2.3298864364624023, "learning_rate": 1.06e-05, "loss": 0.4196, "step": 395 }, { "epoch": 0.13983050847457626, "grad_norm": 4.122328281402588, "learning_rate": 1.05e-05, "loss": 0.3735, "step": 396 }, { "epoch": 0.14018361581920905, "grad_norm": 2.8163208961486816, "learning_rate": 1.04e-05, "loss": 0.4261, "step": 397 }, { "epoch": 0.1405367231638418, "grad_norm": 1.7612152099609375, "learning_rate": 1.03e-05, "loss": 0.4001, "step": 398 }, { "epoch": 0.1408898305084746, "grad_norm": 2.542123317718506, "learning_rate": 1.02e-05, "loss": 0.5853, "step": 399 }, { "epoch": 0.14124293785310735, "grad_norm": 2.013018846511841, "learning_rate": 1.0100000000000002e-05, "loss": 0.344, "step": 400 }, { "epoch": 0.1415960451977401, "grad_norm": 2.6476850509643555, "learning_rate": 1e-05, "loss": 0.5595, "step": 401 }, { "epoch": 0.1419491525423729, "grad_norm": 2.0753297805786133, "learning_rate": 9.900000000000002e-06, "loss": 0.3958, "step": 402 }, { "epoch": 0.14230225988700565, "grad_norm": 1.726521372795105, "learning_rate": 9.800000000000001e-06, "loss": 0.3896, "step": 403 }, { "epoch": 0.1426553672316384, "grad_norm": 2.0191128253936768, "learning_rate": 9.7e-06, "loss": 0.4541, "step": 404 }, { "epoch": 0.1430084745762712, "grad_norm": 2.52801775932312, "learning_rate": 9.600000000000001e-06, "loss": 0.4609, "step": 405 }, { "epoch": 0.14336158192090395, "grad_norm": 2.738588809967041, "learning_rate": 9.5e-06, "loss": 0.4313, "step": 406 }, { "epoch": 0.1437146892655367, "grad_norm": 2.0913681983947754, "learning_rate": 9.4e-06, "loss": 0.4176, "step": 407 }, { "epoch": 0.1440677966101695, "grad_norm": 1.8325021266937256, "learning_rate": 9.3e-06, "loss": 0.3419, "step": 408 }, { "epoch": 0.14442090395480225, "grad_norm": 2.245701551437378, "learning_rate": 9.2e-06, "loss": 0.4799, "step": 409 }, { "epoch": 0.14477401129943504, "grad_norm": 3.9585647583007812, "learning_rate": 9.100000000000001e-06, "loss": 0.6476, "step": 410 }, { "epoch": 0.1451271186440678, "grad_norm": 3.2401275634765625, "learning_rate": 9e-06, "loss": 0.6274, "step": 411 }, { "epoch": 0.14548022598870056, "grad_norm": 2.1975040435791016, "learning_rate": 8.9e-06, "loss": 0.4277, "step": 412 }, { "epoch": 0.14583333333333334, "grad_norm": 1.8795210123062134, "learning_rate": 8.8e-06, "loss": 0.2486, "step": 413 }, { "epoch": 0.1461864406779661, "grad_norm": 2.162196159362793, "learning_rate": 8.7e-06, "loss": 0.3372, "step": 414 }, { "epoch": 0.14653954802259886, "grad_norm": 2.1754472255706787, "learning_rate": 8.599999999999999e-06, "loss": 0.4053, "step": 415 }, { "epoch": 0.14689265536723164, "grad_norm": 2.0052478313446045, "learning_rate": 8.500000000000002e-06, "loss": 0.3775, "step": 416 }, { "epoch": 0.1472457627118644, "grad_norm": 2.1078007221221924, "learning_rate": 8.400000000000001e-06, "loss": 0.5518, "step": 417 }, { "epoch": 0.1475988700564972, "grad_norm": 2.85091495513916, "learning_rate": 8.3e-06, "loss": 0.8025, "step": 418 }, { "epoch": 0.14795197740112995, "grad_norm": 1.9279358386993408, "learning_rate": 8.200000000000001e-06, "loss": 0.4753, "step": 419 }, { "epoch": 0.1483050847457627, "grad_norm": 2.3065125942230225, "learning_rate": 8.1e-06, "loss": 0.3965, "step": 420 }, { "epoch": 0.1486581920903955, "grad_norm": 3.294403076171875, "learning_rate": 8.000000000000001e-06, "loss": 0.6205, "step": 421 }, { "epoch": 0.14901129943502825, "grad_norm": 1.7436306476593018, "learning_rate": 7.9e-06, "loss": 0.239, "step": 422 }, { "epoch": 0.149364406779661, "grad_norm": 1.420858383178711, "learning_rate": 7.8e-06, "loss": 0.2265, "step": 423 }, { "epoch": 0.1497175141242938, "grad_norm": 2.940077066421509, "learning_rate": 7.7e-06, "loss": 0.6006, "step": 424 }, { "epoch": 0.15007062146892655, "grad_norm": 2.1452486515045166, "learning_rate": 7.6e-06, "loss": 0.3333, "step": 425 }, { "epoch": 0.1504237288135593, "grad_norm": 4.092085838317871, "learning_rate": 7.5e-06, "loss": 0.7523, "step": 426 }, { "epoch": 0.1507768361581921, "grad_norm": 1.8198660612106323, "learning_rate": 7.4e-06, "loss": 0.3283, "step": 427 }, { "epoch": 0.15112994350282485, "grad_norm": 1.3757944107055664, "learning_rate": 7.2999999999999996e-06, "loss": 0.2199, "step": 428 }, { "epoch": 0.15148305084745764, "grad_norm": 2.0032095909118652, "learning_rate": 7.2e-06, "loss": 0.3994, "step": 429 }, { "epoch": 0.1518361581920904, "grad_norm": 2.9988086223602295, "learning_rate": 7.1e-06, "loss": 0.3558, "step": 430 }, { "epoch": 0.15218926553672316, "grad_norm": 2.7509193420410156, "learning_rate": 7.000000000000001e-06, "loss": 0.5825, "step": 431 }, { "epoch": 0.15254237288135594, "grad_norm": 2.523036003112793, "learning_rate": 6.900000000000001e-06, "loss": 0.484, "step": 432 }, { "epoch": 0.1528954802259887, "grad_norm": 1.8027594089508057, "learning_rate": 6.800000000000001e-06, "loss": 0.382, "step": 433 }, { "epoch": 0.15324858757062146, "grad_norm": 2.55008602142334, "learning_rate": 6.700000000000001e-06, "loss": 0.4458, "step": 434 }, { "epoch": 0.15360169491525424, "grad_norm": 2.159984827041626, "learning_rate": 6.6e-06, "loss": 0.3597, "step": 435 }, { "epoch": 0.153954802259887, "grad_norm": 2.2070958614349365, "learning_rate": 6.5000000000000004e-06, "loss": 0.483, "step": 436 }, { "epoch": 0.15430790960451976, "grad_norm": 2.283949851989746, "learning_rate": 6.4000000000000006e-06, "loss": 0.3999, "step": 437 }, { "epoch": 0.15466101694915255, "grad_norm": 2.086303472518921, "learning_rate": 6.300000000000001e-06, "loss": 0.4877, "step": 438 }, { "epoch": 0.1550141242937853, "grad_norm": 2.2829809188842773, "learning_rate": 6.2e-06, "loss": 0.4839, "step": 439 }, { "epoch": 0.1553672316384181, "grad_norm": 2.9696552753448486, "learning_rate": 6.1e-06, "loss": 0.5502, "step": 440 }, { "epoch": 0.15572033898305085, "grad_norm": 2.7994322776794434, "learning_rate": 6e-06, "loss": 0.5695, "step": 441 }, { "epoch": 0.1560734463276836, "grad_norm": 1.9340434074401855, "learning_rate": 5.9e-06, "loss": 0.3296, "step": 442 }, { "epoch": 0.1564265536723164, "grad_norm": 2.5606064796447754, "learning_rate": 5.8e-06, "loss": 0.411, "step": 443 }, { "epoch": 0.15677966101694915, "grad_norm": 2.6650333404541016, "learning_rate": 5.7000000000000005e-06, "loss": 0.3179, "step": 444 }, { "epoch": 0.1571327683615819, "grad_norm": 4.912551403045654, "learning_rate": 5.600000000000001e-06, "loss": 0.8535, "step": 445 }, { "epoch": 0.1574858757062147, "grad_norm": 2.2671737670898438, "learning_rate": 5.500000000000001e-06, "loss": 0.3959, "step": 446 }, { "epoch": 0.15783898305084745, "grad_norm": 2.4728825092315674, "learning_rate": 5.4e-06, "loss": 0.3118, "step": 447 }, { "epoch": 0.15819209039548024, "grad_norm": 1.7785649299621582, "learning_rate": 5.3e-06, "loss": 0.228, "step": 448 }, { "epoch": 0.158545197740113, "grad_norm": 2.288245916366577, "learning_rate": 5.2e-06, "loss": 0.3586, "step": 449 }, { "epoch": 0.15889830508474576, "grad_norm": 1.821871042251587, "learning_rate": 5.1e-06, "loss": 0.362, "step": 450 }, { "epoch": 0.15925141242937854, "grad_norm": 1.9763644933700562, "learning_rate": 5e-06, "loss": 0.3057, "step": 451 }, { "epoch": 0.1596045197740113, "grad_norm": 1.8915050029754639, "learning_rate": 4.9000000000000005e-06, "loss": 0.3815, "step": 452 }, { "epoch": 0.15995762711864406, "grad_norm": 1.787156343460083, "learning_rate": 4.800000000000001e-06, "loss": 0.2223, "step": 453 }, { "epoch": 0.16031073446327684, "grad_norm": 2.0852158069610596, "learning_rate": 4.7e-06, "loss": 0.4158, "step": 454 }, { "epoch": 0.1606638418079096, "grad_norm": 1.9223723411560059, "learning_rate": 4.6e-06, "loss": 0.3043, "step": 455 }, { "epoch": 0.16101694915254236, "grad_norm": 1.8598310947418213, "learning_rate": 4.5e-06, "loss": 0.3886, "step": 456 }, { "epoch": 0.16137005649717515, "grad_norm": 2.242039918899536, "learning_rate": 4.4e-06, "loss": 0.4703, "step": 457 }, { "epoch": 0.1617231638418079, "grad_norm": 1.7315804958343506, "learning_rate": 4.2999999999999995e-06, "loss": 0.2178, "step": 458 }, { "epoch": 0.1620762711864407, "grad_norm": 2.204817295074463, "learning_rate": 4.2000000000000004e-06, "loss": 0.3898, "step": 459 }, { "epoch": 0.16242937853107345, "grad_norm": 2.1929430961608887, "learning_rate": 4.1000000000000006e-06, "loss": 0.3549, "step": 460 }, { "epoch": 0.1627824858757062, "grad_norm": 1.7955076694488525, "learning_rate": 4.000000000000001e-06, "loss": 0.3815, "step": 461 }, { "epoch": 0.163135593220339, "grad_norm": 4.15421724319458, "learning_rate": 3.9e-06, "loss": 0.6433, "step": 462 }, { "epoch": 0.16348870056497175, "grad_norm": 1.9674700498580933, "learning_rate": 3.8e-06, "loss": 0.3425, "step": 463 }, { "epoch": 0.1638418079096045, "grad_norm": 1.909616231918335, "learning_rate": 3.7e-06, "loss": 0.343, "step": 464 }, { "epoch": 0.1641949152542373, "grad_norm": 2.229336977005005, "learning_rate": 3.6e-06, "loss": 0.4534, "step": 465 }, { "epoch": 0.16454802259887005, "grad_norm": 1.3924556970596313, "learning_rate": 3.5000000000000004e-06, "loss": 0.2077, "step": 466 }, { "epoch": 0.1649011299435028, "grad_norm": 1.3659465312957764, "learning_rate": 3.4000000000000005e-06, "loss": 0.2068, "step": 467 }, { "epoch": 0.1652542372881356, "grad_norm": 1.867235779762268, "learning_rate": 3.3e-06, "loss": 0.2963, "step": 468 }, { "epoch": 0.16560734463276836, "grad_norm": 1.332623839378357, "learning_rate": 3.2000000000000003e-06, "loss": 0.2054, "step": 469 }, { "epoch": 0.16596045197740114, "grad_norm": 1.6542198657989502, "learning_rate": 3.1e-06, "loss": 0.2103, "step": 470 }, { "epoch": 0.1663135593220339, "grad_norm": 2.120319128036499, "learning_rate": 3e-06, "loss": 0.4215, "step": 471 }, { "epoch": 0.16666666666666666, "grad_norm": 1.8779698610305786, "learning_rate": 2.9e-06, "loss": 0.298, "step": 472 }, { "epoch": 0.16701977401129944, "grad_norm": 2.830299139022827, "learning_rate": 2.8000000000000003e-06, "loss": 0.542, "step": 473 }, { "epoch": 0.1673728813559322, "grad_norm": 2.909973621368408, "learning_rate": 2.7e-06, "loss": 0.4126, "step": 474 }, { "epoch": 0.16772598870056496, "grad_norm": 4.498685359954834, "learning_rate": 2.6e-06, "loss": 0.5541, "step": 475 }, { "epoch": 0.16807909604519775, "grad_norm": 2.452082395553589, "learning_rate": 2.5e-06, "loss": 0.3653, "step": 476 }, { "epoch": 0.1684322033898305, "grad_norm": 1.9502593278884888, "learning_rate": 2.4000000000000003e-06, "loss": 0.3149, "step": 477 }, { "epoch": 0.1687853107344633, "grad_norm": 2.2951347827911377, "learning_rate": 2.3e-06, "loss": 0.3794, "step": 478 }, { "epoch": 0.16913841807909605, "grad_norm": 3.0120177268981934, "learning_rate": 2.2e-06, "loss": 0.5245, "step": 479 }, { "epoch": 0.1694915254237288, "grad_norm": 4.143613815307617, "learning_rate": 2.1000000000000002e-06, "loss": 0.6444, "step": 480 }, { "epoch": 0.1698446327683616, "grad_norm": 1.838753342628479, "learning_rate": 2.0000000000000003e-06, "loss": 0.3379, "step": 481 }, { "epoch": 0.17019774011299435, "grad_norm": 1.9792391061782837, "learning_rate": 1.9e-06, "loss": 0.3329, "step": 482 }, { "epoch": 0.1705508474576271, "grad_norm": 1.8502469062805176, "learning_rate": 1.8e-06, "loss": 0.3703, "step": 483 }, { "epoch": 0.1709039548022599, "grad_norm": 1.618812918663025, "learning_rate": 1.7000000000000002e-06, "loss": 0.2051, "step": 484 }, { "epoch": 0.17125706214689265, "grad_norm": 2.346790075302124, "learning_rate": 1.6000000000000001e-06, "loss": 0.49, "step": 485 }, { "epoch": 0.1716101694915254, "grad_norm": 1.8763465881347656, "learning_rate": 1.5e-06, "loss": 0.2948, "step": 486 }, { "epoch": 0.1719632768361582, "grad_norm": 1.7063709497451782, "learning_rate": 1.4000000000000001e-06, "loss": 0.3513, "step": 487 }, { "epoch": 0.17231638418079095, "grad_norm": 2.963041067123413, "learning_rate": 1.3e-06, "loss": 0.6783, "step": 488 }, { "epoch": 0.17266949152542374, "grad_norm": 2.9806196689605713, "learning_rate": 1.2000000000000002e-06, "loss": 0.5123, "step": 489 }, { "epoch": 0.1730225988700565, "grad_norm": 1.6302564144134521, "learning_rate": 1.1e-06, "loss": 0.204, "step": 490 }, { "epoch": 0.17337570621468926, "grad_norm": 1.9233272075653076, "learning_rate": 1.0000000000000002e-06, "loss": 0.3037, "step": 491 }, { "epoch": 0.17372881355932204, "grad_norm": 1.8432778120040894, "learning_rate": 9e-07, "loss": 0.3368, "step": 492 }, { "epoch": 0.1740819209039548, "grad_norm": 2.5116734504699707, "learning_rate": 8.000000000000001e-07, "loss": 0.5263, "step": 493 }, { "epoch": 0.17443502824858756, "grad_norm": 3.0080406665802, "learning_rate": 7.000000000000001e-07, "loss": 0.3415, "step": 494 }, { "epoch": 0.17478813559322035, "grad_norm": 1.982960820198059, "learning_rate": 6.000000000000001e-07, "loss": 0.459, "step": 495 }, { "epoch": 0.1751412429378531, "grad_norm": 2.259504795074463, "learning_rate": 5.000000000000001e-07, "loss": 0.4088, "step": 496 }, { "epoch": 0.1754943502824859, "grad_norm": 1.937008261680603, "learning_rate": 4.0000000000000003e-07, "loss": 0.2903, "step": 497 }, { "epoch": 0.17584745762711865, "grad_norm": 1.8799011707305908, "learning_rate": 3.0000000000000004e-07, "loss": 0.2905, "step": 498 }, { "epoch": 0.1762005649717514, "grad_norm": 2.9519193172454834, "learning_rate": 2.0000000000000002e-07, "loss": 0.525, "step": 499 }, { "epoch": 0.1765536723163842, "grad_norm": 1.2929198741912842, "learning_rate": 1.0000000000000001e-07, "loss": 0.1987, "step": 500 } ], "logging_steps": 1, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.099784700491894e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }