{ "best_metric": 1.0619021654129028, "best_model_checkpoint": "square_run_second_vote/checkpoint-464", "epoch": 30.0, "eval_steps": 500, "global_step": 1740, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.034482758620689655, "grad_norm": 10.100388526916504, "learning_rate": 1.1494252873563219e-06, "loss": 1.9216, "step": 2 }, { "epoch": 0.06896551724137931, "grad_norm": 15.660673141479492, "learning_rate": 2.2988505747126437e-06, "loss": 1.9891, "step": 4 }, { "epoch": 0.10344827586206896, "grad_norm": 13.529168128967285, "learning_rate": 3.448275862068966e-06, "loss": 2.0535, "step": 6 }, { "epoch": 0.13793103448275862, "grad_norm": 12.513371467590332, "learning_rate": 4.5977011494252875e-06, "loss": 2.0399, "step": 8 }, { "epoch": 0.1724137931034483, "grad_norm": 12.855453491210938, "learning_rate": 5.747126436781609e-06, "loss": 1.9636, "step": 10 }, { "epoch": 0.20689655172413793, "grad_norm": 9.419121742248535, "learning_rate": 6.896551724137932e-06, "loss": 1.7951, "step": 12 }, { "epoch": 0.2413793103448276, "grad_norm": 8.58608627319336, "learning_rate": 8.045977011494253e-06, "loss": 1.938, "step": 14 }, { "epoch": 0.27586206896551724, "grad_norm": 15.212312698364258, "learning_rate": 9.195402298850575e-06, "loss": 1.947, "step": 16 }, { "epoch": 0.3103448275862069, "grad_norm": 10.53651237487793, "learning_rate": 1.0344827586206897e-05, "loss": 2.001, "step": 18 }, { "epoch": 0.3448275862068966, "grad_norm": 10.844449043273926, "learning_rate": 1.1494252873563218e-05, "loss": 1.6643, "step": 20 }, { "epoch": 0.3793103448275862, "grad_norm": 12.122403144836426, "learning_rate": 1.2643678160919542e-05, "loss": 1.7857, "step": 22 }, { "epoch": 0.41379310344827586, "grad_norm": 10.027009963989258, "learning_rate": 1.3793103448275863e-05, "loss": 2.0627, "step": 24 }, { "epoch": 0.4482758620689655, "grad_norm": 10.017681121826172, "learning_rate": 1.4942528735632185e-05, "loss": 1.8997, "step": 26 }, { "epoch": 0.4827586206896552, "grad_norm": 11.976802825927734, "learning_rate": 1.6091954022988507e-05, "loss": 1.9658, "step": 28 }, { "epoch": 0.5172413793103449, "grad_norm": 17.975006103515625, "learning_rate": 1.7241379310344828e-05, "loss": 2.1084, "step": 30 }, { "epoch": 0.5517241379310345, "grad_norm": 10.179977416992188, "learning_rate": 1.839080459770115e-05, "loss": 1.8625, "step": 32 }, { "epoch": 0.5862068965517241, "grad_norm": 12.881868362426758, "learning_rate": 1.9540229885057475e-05, "loss": 2.0918, "step": 34 }, { "epoch": 0.6206896551724138, "grad_norm": 11.166668891906738, "learning_rate": 2.0689655172413793e-05, "loss": 1.7992, "step": 36 }, { "epoch": 0.6551724137931034, "grad_norm": 9.509991645812988, "learning_rate": 2.183908045977012e-05, "loss": 1.7821, "step": 38 }, { "epoch": 0.6896551724137931, "grad_norm": 9.292908668518066, "learning_rate": 2.2988505747126437e-05, "loss": 1.8435, "step": 40 }, { "epoch": 0.7241379310344828, "grad_norm": 9.034067153930664, "learning_rate": 2.413793103448276e-05, "loss": 1.7203, "step": 42 }, { "epoch": 0.7586206896551724, "grad_norm": 10.527994155883789, "learning_rate": 2.5287356321839083e-05, "loss": 2.0411, "step": 44 }, { "epoch": 0.7931034482758621, "grad_norm": 12.680684089660645, "learning_rate": 2.6436781609195405e-05, "loss": 1.7733, "step": 46 }, { "epoch": 0.8275862068965517, "grad_norm": 11.482306480407715, "learning_rate": 2.7586206896551727e-05, "loss": 1.9162, "step": 48 }, { "epoch": 0.8620689655172413, "grad_norm": 9.575610160827637, "learning_rate": 2.8735632183908045e-05, "loss": 1.9413, "step": 50 }, { "epoch": 0.896551724137931, "grad_norm": 9.212911605834961, "learning_rate": 2.988505747126437e-05, "loss": 1.6267, "step": 52 }, { "epoch": 0.9310344827586207, "grad_norm": 10.29886245727539, "learning_rate": 3.103448275862069e-05, "loss": 2.0705, "step": 54 }, { "epoch": 0.9655172413793104, "grad_norm": 10.353363990783691, "learning_rate": 3.218390804597701e-05, "loss": 1.7528, "step": 56 }, { "epoch": 1.0, "grad_norm": 10.195358276367188, "learning_rate": 3.3333333333333335e-05, "loss": 1.8754, "step": 58 }, { "epoch": 1.0, "eval_accuracy": 0.2803030303030303, "eval_f1_macro": 0.13852914120300855, "eval_f1_micro": 0.2803030303030303, "eval_f1_weighted": 0.17218993609140762, "eval_loss": 1.7960734367370605, "eval_precision_macro": 0.14259740259740258, "eval_precision_micro": 0.2803030303030303, "eval_precision_weighted": 0.16029384756657483, "eval_recall_macro": 0.20982142857142858, "eval_recall_micro": 0.2803030303030303, "eval_recall_weighted": 0.2803030303030303, "eval_runtime": 2.0654, "eval_samples_per_second": 63.909, "eval_steps_per_second": 8.231, "step": 58 }, { "epoch": 1.0344827586206897, "grad_norm": 6.763228416442871, "learning_rate": 3.4482758620689657e-05, "loss": 1.556, "step": 60 }, { "epoch": 1.0689655172413792, "grad_norm": 12.80778694152832, "learning_rate": 3.563218390804598e-05, "loss": 1.8112, "step": 62 }, { "epoch": 1.103448275862069, "grad_norm": 5.85200309753418, "learning_rate": 3.67816091954023e-05, "loss": 1.6181, "step": 64 }, { "epoch": 1.1379310344827587, "grad_norm": 13.358109474182129, "learning_rate": 3.793103448275862e-05, "loss": 1.5691, "step": 66 }, { "epoch": 1.1724137931034484, "grad_norm": 13.59086799621582, "learning_rate": 3.908045977011495e-05, "loss": 2.028, "step": 68 }, { "epoch": 1.206896551724138, "grad_norm": 12.442203521728516, "learning_rate": 4.0229885057471265e-05, "loss": 1.8991, "step": 70 }, { "epoch": 1.2413793103448276, "grad_norm": 10.374919891357422, "learning_rate": 4.1379310344827587e-05, "loss": 1.7926, "step": 72 }, { "epoch": 1.2758620689655173, "grad_norm": 8.41528034210205, "learning_rate": 4.252873563218391e-05, "loss": 1.9024, "step": 74 }, { "epoch": 1.3103448275862069, "grad_norm": 8.301165580749512, "learning_rate": 4.367816091954024e-05, "loss": 1.5357, "step": 76 }, { "epoch": 1.3448275862068966, "grad_norm": 10.928062438964844, "learning_rate": 4.482758620689655e-05, "loss": 1.7872, "step": 78 }, { "epoch": 1.3793103448275863, "grad_norm": 12.0917329788208, "learning_rate": 4.597701149425287e-05, "loss": 1.9049, "step": 80 }, { "epoch": 1.4137931034482758, "grad_norm": 10.680750846862793, "learning_rate": 4.7126436781609195e-05, "loss": 1.818, "step": 82 }, { "epoch": 1.4482758620689655, "grad_norm": 10.11230754852295, "learning_rate": 4.827586206896552e-05, "loss": 1.7855, "step": 84 }, { "epoch": 1.4827586206896552, "grad_norm": 15.761420249938965, "learning_rate": 4.9425287356321845e-05, "loss": 1.7779, "step": 86 }, { "epoch": 1.5172413793103448, "grad_norm": 15.814207077026367, "learning_rate": 5.057471264367817e-05, "loss": 1.6615, "step": 88 }, { "epoch": 1.5517241379310345, "grad_norm": 10.473943710327148, "learning_rate": 5.172413793103449e-05, "loss": 1.7123, "step": 90 }, { "epoch": 1.5862068965517242, "grad_norm": 9.527724266052246, "learning_rate": 5.287356321839081e-05, "loss": 1.5118, "step": 92 }, { "epoch": 1.6206896551724137, "grad_norm": 13.386604309082031, "learning_rate": 5.402298850574713e-05, "loss": 2.0467, "step": 94 }, { "epoch": 1.6551724137931034, "grad_norm": 11.433362007141113, "learning_rate": 5.517241379310345e-05, "loss": 1.7908, "step": 96 }, { "epoch": 1.6896551724137931, "grad_norm": 10.231264114379883, "learning_rate": 5.632183908045977e-05, "loss": 2.0643, "step": 98 }, { "epoch": 1.7241379310344827, "grad_norm": 10.50631046295166, "learning_rate": 5.747126436781609e-05, "loss": 1.6573, "step": 100 }, { "epoch": 1.7586206896551724, "grad_norm": 10.044520378112793, "learning_rate": 5.862068965517241e-05, "loss": 1.6, "step": 102 }, { "epoch": 1.793103448275862, "grad_norm": 11.720151901245117, "learning_rate": 5.977011494252874e-05, "loss": 1.9807, "step": 104 }, { "epoch": 1.8275862068965516, "grad_norm": 8.176350593566895, "learning_rate": 6.091954022988506e-05, "loss": 1.772, "step": 106 }, { "epoch": 1.8620689655172413, "grad_norm": 10.523119926452637, "learning_rate": 6.206896551724138e-05, "loss": 1.5979, "step": 108 }, { "epoch": 1.896551724137931, "grad_norm": 12.917177200317383, "learning_rate": 6.32183908045977e-05, "loss": 1.9565, "step": 110 }, { "epoch": 1.9310344827586206, "grad_norm": 9.43907356262207, "learning_rate": 6.436781609195403e-05, "loss": 1.6814, "step": 112 }, { "epoch": 1.9655172413793105, "grad_norm": 7.691906452178955, "learning_rate": 6.551724137931034e-05, "loss": 1.8594, "step": 114 }, { "epoch": 2.0, "grad_norm": 12.988773345947266, "learning_rate": 6.666666666666667e-05, "loss": 2.0246, "step": 116 }, { "epoch": 2.0, "eval_accuracy": 0.3106060606060606, "eval_f1_macro": 0.22363870840750824, "eval_f1_micro": 0.3106060606060606, "eval_f1_weighted": 0.2483673419087291, "eval_loss": 2.0137689113616943, "eval_precision_macro": 0.2557566049675684, "eval_precision_micro": 0.3106060606060606, "eval_precision_weighted": 0.2691759790597, "eval_recall_macro": 0.2841635338345864, "eval_recall_micro": 0.3106060606060606, "eval_recall_weighted": 0.3106060606060606, "eval_runtime": 1.9498, "eval_samples_per_second": 67.698, "eval_steps_per_second": 8.719, "step": 116 }, { "epoch": 2.0344827586206895, "grad_norm": 10.840887069702148, "learning_rate": 6.781609195402298e-05, "loss": 1.7498, "step": 118 }, { "epoch": 2.0689655172413794, "grad_norm": 11.36091423034668, "learning_rate": 6.896551724137931e-05, "loss": 1.7945, "step": 120 }, { "epoch": 2.103448275862069, "grad_norm": 9.655730247497559, "learning_rate": 7.011494252873564e-05, "loss": 1.6088, "step": 122 }, { "epoch": 2.1379310344827585, "grad_norm": 9.597773551940918, "learning_rate": 7.126436781609196e-05, "loss": 1.6544, "step": 124 }, { "epoch": 2.1724137931034484, "grad_norm": 8.65636157989502, "learning_rate": 7.241379310344828e-05, "loss": 1.4854, "step": 126 }, { "epoch": 2.206896551724138, "grad_norm": 8.050541877746582, "learning_rate": 7.35632183908046e-05, "loss": 1.3232, "step": 128 }, { "epoch": 2.2413793103448274, "grad_norm": 11.399789810180664, "learning_rate": 7.471264367816091e-05, "loss": 1.6108, "step": 130 }, { "epoch": 2.2758620689655173, "grad_norm": 10.911064147949219, "learning_rate": 7.586206896551724e-05, "loss": 1.3609, "step": 132 }, { "epoch": 2.310344827586207, "grad_norm": 10.229703903198242, "learning_rate": 7.701149425287356e-05, "loss": 1.4551, "step": 134 }, { "epoch": 2.344827586206897, "grad_norm": 10.077006340026855, "learning_rate": 7.81609195402299e-05, "loss": 1.3409, "step": 136 }, { "epoch": 2.3793103448275863, "grad_norm": 12.368110656738281, "learning_rate": 7.931034482758621e-05, "loss": 1.8165, "step": 138 }, { "epoch": 2.413793103448276, "grad_norm": 16.638439178466797, "learning_rate": 8.045977011494253e-05, "loss": 1.6881, "step": 140 }, { "epoch": 2.4482758620689653, "grad_norm": 10.63574504852295, "learning_rate": 8.160919540229886e-05, "loss": 1.3659, "step": 142 }, { "epoch": 2.4827586206896552, "grad_norm": 12.455951690673828, "learning_rate": 8.275862068965517e-05, "loss": 1.3084, "step": 144 }, { "epoch": 2.5172413793103448, "grad_norm": 7.663853168487549, "learning_rate": 8.39080459770115e-05, "loss": 1.3424, "step": 146 }, { "epoch": 2.5517241379310347, "grad_norm": 8.313244819641113, "learning_rate": 8.505747126436782e-05, "loss": 1.2946, "step": 148 }, { "epoch": 2.586206896551724, "grad_norm": 9.524584770202637, "learning_rate": 8.620689655172413e-05, "loss": 1.1958, "step": 150 }, { "epoch": 2.6206896551724137, "grad_norm": 12.151946067810059, "learning_rate": 8.735632183908047e-05, "loss": 1.6999, "step": 152 }, { "epoch": 2.655172413793103, "grad_norm": 9.74065113067627, "learning_rate": 8.850574712643679e-05, "loss": 1.5755, "step": 154 }, { "epoch": 2.689655172413793, "grad_norm": 14.421919822692871, "learning_rate": 8.96551724137931e-05, "loss": 1.428, "step": 156 }, { "epoch": 2.7241379310344827, "grad_norm": 11.11561393737793, "learning_rate": 9.080459770114943e-05, "loss": 1.4898, "step": 158 }, { "epoch": 2.7586206896551726, "grad_norm": 13.893324851989746, "learning_rate": 9.195402298850575e-05, "loss": 1.901, "step": 160 }, { "epoch": 2.793103448275862, "grad_norm": 8.19894027709961, "learning_rate": 9.310344827586207e-05, "loss": 1.4159, "step": 162 }, { "epoch": 2.8275862068965516, "grad_norm": 10.198213577270508, "learning_rate": 9.425287356321839e-05, "loss": 1.6808, "step": 164 }, { "epoch": 2.862068965517241, "grad_norm": 7.003215312957764, "learning_rate": 9.540229885057472e-05, "loss": 1.859, "step": 166 }, { "epoch": 2.896551724137931, "grad_norm": 8.648194313049316, "learning_rate": 9.655172413793105e-05, "loss": 1.5442, "step": 168 }, { "epoch": 2.9310344827586206, "grad_norm": 9.167272567749023, "learning_rate": 9.770114942528736e-05, "loss": 1.6219, "step": 170 }, { "epoch": 2.9655172413793105, "grad_norm": 13.052528381347656, "learning_rate": 9.885057471264369e-05, "loss": 1.637, "step": 172 }, { "epoch": 3.0, "grad_norm": 10.723739624023438, "learning_rate": 0.0001, "loss": 1.6189, "step": 174 }, { "epoch": 3.0, "eval_accuracy": 0.38636363636363635, "eval_f1_macro": 0.244379876244283, "eval_f1_micro": 0.38636363636363635, "eval_f1_weighted": 0.3195200593505679, "eval_loss": 1.5038808584213257, "eval_precision_macro": 0.2633030990173847, "eval_precision_micro": 0.38636363636363635, "eval_precision_weighted": 0.33009660092993426, "eval_recall_macro": 0.28469611528822053, "eval_recall_micro": 0.38636363636363635, "eval_recall_weighted": 0.38636363636363635, "eval_runtime": 1.953, "eval_samples_per_second": 67.59, "eval_steps_per_second": 8.705, "step": 174 }, { "epoch": 3.0344827586206895, "grad_norm": 7.145821571350098, "learning_rate": 9.987228607918264e-05, "loss": 1.3525, "step": 176 }, { "epoch": 3.0689655172413794, "grad_norm": 8.73121166229248, "learning_rate": 9.974457215836527e-05, "loss": 1.3759, "step": 178 }, { "epoch": 3.103448275862069, "grad_norm": 11.502145767211914, "learning_rate": 9.96168582375479e-05, "loss": 1.3372, "step": 180 }, { "epoch": 3.1379310344827585, "grad_norm": 8.244510650634766, "learning_rate": 9.948914431673053e-05, "loss": 1.5859, "step": 182 }, { "epoch": 3.1724137931034484, "grad_norm": 9.313932418823242, "learning_rate": 9.936143039591316e-05, "loss": 1.4753, "step": 184 }, { "epoch": 3.206896551724138, "grad_norm": 13.148027420043945, "learning_rate": 9.92337164750958e-05, "loss": 1.5913, "step": 186 }, { "epoch": 3.2413793103448274, "grad_norm": 14.681943893432617, "learning_rate": 9.910600255427843e-05, "loss": 1.4892, "step": 188 }, { "epoch": 3.2758620689655173, "grad_norm": 13.758460998535156, "learning_rate": 9.897828863346104e-05, "loss": 1.333, "step": 190 }, { "epoch": 3.310344827586207, "grad_norm": 11.166180610656738, "learning_rate": 9.885057471264369e-05, "loss": 1.5819, "step": 192 }, { "epoch": 3.344827586206897, "grad_norm": 9.059310913085938, "learning_rate": 9.872286079182631e-05, "loss": 0.9963, "step": 194 }, { "epoch": 3.3793103448275863, "grad_norm": 9.729276657104492, "learning_rate": 9.859514687100895e-05, "loss": 1.8236, "step": 196 }, { "epoch": 3.413793103448276, "grad_norm": 11.96312427520752, "learning_rate": 9.846743295019157e-05, "loss": 1.4265, "step": 198 }, { "epoch": 3.4482758620689653, "grad_norm": 11.733696937561035, "learning_rate": 9.833971902937422e-05, "loss": 1.1957, "step": 200 }, { "epoch": 3.4827586206896552, "grad_norm": 19.109695434570312, "learning_rate": 9.821200510855683e-05, "loss": 2.2991, "step": 202 }, { "epoch": 3.5172413793103448, "grad_norm": 11.54105281829834, "learning_rate": 9.808429118773947e-05, "loss": 1.5837, "step": 204 }, { "epoch": 3.5517241379310347, "grad_norm": 14.453530311584473, "learning_rate": 9.79565772669221e-05, "loss": 1.4211, "step": 206 }, { "epoch": 3.586206896551724, "grad_norm": 12.695812225341797, "learning_rate": 9.782886334610473e-05, "loss": 1.0612, "step": 208 }, { "epoch": 3.6206896551724137, "grad_norm": 9.610103607177734, "learning_rate": 9.770114942528736e-05, "loss": 1.2058, "step": 210 }, { "epoch": 3.655172413793103, "grad_norm": 12.852982521057129, "learning_rate": 9.757343550446999e-05, "loss": 1.735, "step": 212 }, { "epoch": 3.689655172413793, "grad_norm": 10.032821655273438, "learning_rate": 9.744572158365262e-05, "loss": 1.6214, "step": 214 }, { "epoch": 3.7241379310344827, "grad_norm": 13.28847885131836, "learning_rate": 9.731800766283526e-05, "loss": 1.2198, "step": 216 }, { "epoch": 3.7586206896551726, "grad_norm": 9.55737590789795, "learning_rate": 9.719029374201787e-05, "loss": 1.3811, "step": 218 }, { "epoch": 3.793103448275862, "grad_norm": 12.440496444702148, "learning_rate": 9.706257982120052e-05, "loss": 1.2533, "step": 220 }, { "epoch": 3.8275862068965516, "grad_norm": 9.515503883361816, "learning_rate": 9.693486590038314e-05, "loss": 1.1579, "step": 222 }, { "epoch": 3.862068965517241, "grad_norm": 14.026083946228027, "learning_rate": 9.680715197956578e-05, "loss": 1.7226, "step": 224 }, { "epoch": 3.896551724137931, "grad_norm": 10.697053909301758, "learning_rate": 9.66794380587484e-05, "loss": 1.1127, "step": 226 }, { "epoch": 3.9310344827586206, "grad_norm": 11.028640747070312, "learning_rate": 9.655172413793105e-05, "loss": 1.3266, "step": 228 }, { "epoch": 3.9655172413793105, "grad_norm": 15.626441955566406, "learning_rate": 9.642401021711366e-05, "loss": 1.9315, "step": 230 }, { "epoch": 4.0, "grad_norm": 13.396081924438477, "learning_rate": 9.62962962962963e-05, "loss": 1.3445, "step": 232 }, { "epoch": 4.0, "eval_accuracy": 0.4393939393939394, "eval_f1_macro": 0.3286744239291642, "eval_f1_micro": 0.4393939393939394, "eval_f1_weighted": 0.38656078267083965, "eval_loss": 1.3982452154159546, "eval_precision_macro": 0.31856682769726247, "eval_precision_micro": 0.4393939393939394, "eval_precision_weighted": 0.37300139071878197, "eval_recall_macro": 0.3695958646616541, "eval_recall_micro": 0.4393939393939394, "eval_recall_weighted": 0.4393939393939394, "eval_runtime": 1.9654, "eval_samples_per_second": 67.161, "eval_steps_per_second": 8.65, "step": 232 }, { "epoch": 4.0344827586206895, "grad_norm": 9.478927612304688, "learning_rate": 9.616858237547893e-05, "loss": 1.3661, "step": 234 }, { "epoch": 4.068965517241379, "grad_norm": 9.680974006652832, "learning_rate": 9.604086845466156e-05, "loss": 1.0935, "step": 236 }, { "epoch": 4.103448275862069, "grad_norm": 7.175812244415283, "learning_rate": 9.591315453384419e-05, "loss": 1.1414, "step": 238 }, { "epoch": 4.137931034482759, "grad_norm": 9.398384094238281, "learning_rate": 9.578544061302682e-05, "loss": 1.1399, "step": 240 }, { "epoch": 4.172413793103448, "grad_norm": 6.083748817443848, "learning_rate": 9.565772669220945e-05, "loss": 0.8365, "step": 242 }, { "epoch": 4.206896551724138, "grad_norm": 7.871669292449951, "learning_rate": 9.553001277139209e-05, "loss": 0.747, "step": 244 }, { "epoch": 4.241379310344827, "grad_norm": 12.75833797454834, "learning_rate": 9.540229885057472e-05, "loss": 1.1442, "step": 246 }, { "epoch": 4.275862068965517, "grad_norm": 9.026577949523926, "learning_rate": 9.527458492975735e-05, "loss": 1.1016, "step": 248 }, { "epoch": 4.310344827586207, "grad_norm": 11.575862884521484, "learning_rate": 9.514687100893998e-05, "loss": 1.6752, "step": 250 }, { "epoch": 4.344827586206897, "grad_norm": 11.908188819885254, "learning_rate": 9.501915708812261e-05, "loss": 1.6481, "step": 252 }, { "epoch": 4.379310344827586, "grad_norm": 11.563265800476074, "learning_rate": 9.489144316730524e-05, "loss": 1.2965, "step": 254 }, { "epoch": 4.413793103448276, "grad_norm": 8.027771949768066, "learning_rate": 9.476372924648788e-05, "loss": 1.1308, "step": 256 }, { "epoch": 4.448275862068965, "grad_norm": 8.491975784301758, "learning_rate": 9.463601532567051e-05, "loss": 0.9251, "step": 258 }, { "epoch": 4.482758620689655, "grad_norm": 13.935683250427246, "learning_rate": 9.450830140485314e-05, "loss": 1.7397, "step": 260 }, { "epoch": 4.517241379310345, "grad_norm": 14.531791687011719, "learning_rate": 9.438058748403577e-05, "loss": 1.67, "step": 262 }, { "epoch": 4.551724137931035, "grad_norm": 9.794015884399414, "learning_rate": 9.425287356321839e-05, "loss": 1.0594, "step": 264 }, { "epoch": 4.586206896551724, "grad_norm": 9.404264450073242, "learning_rate": 9.412515964240103e-05, "loss": 1.1629, "step": 266 }, { "epoch": 4.620689655172414, "grad_norm": 11.694725036621094, "learning_rate": 9.399744572158365e-05, "loss": 1.2181, "step": 268 }, { "epoch": 4.655172413793103, "grad_norm": 12.964648246765137, "learning_rate": 9.38697318007663e-05, "loss": 1.2802, "step": 270 }, { "epoch": 4.689655172413794, "grad_norm": 9.874382972717285, "learning_rate": 9.374201787994892e-05, "loss": 1.124, "step": 272 }, { "epoch": 4.724137931034483, "grad_norm": 11.962569236755371, "learning_rate": 9.361430395913156e-05, "loss": 1.6443, "step": 274 }, { "epoch": 4.758620689655173, "grad_norm": 9.690713882446289, "learning_rate": 9.348659003831418e-05, "loss": 1.6067, "step": 276 }, { "epoch": 4.793103448275862, "grad_norm": 11.790706634521484, "learning_rate": 9.335887611749681e-05, "loss": 1.29, "step": 278 }, { "epoch": 4.827586206896552, "grad_norm": 8.600451469421387, "learning_rate": 9.323116219667944e-05, "loss": 1.696, "step": 280 }, { "epoch": 4.862068965517241, "grad_norm": 7.481589317321777, "learning_rate": 9.310344827586207e-05, "loss": 1.1751, "step": 282 }, { "epoch": 4.896551724137931, "grad_norm": 11.036040306091309, "learning_rate": 9.29757343550447e-05, "loss": 0.9911, "step": 284 }, { "epoch": 4.931034482758621, "grad_norm": 10.061944961547852, "learning_rate": 9.284802043422734e-05, "loss": 1.4282, "step": 286 }, { "epoch": 4.9655172413793105, "grad_norm": 8.946602821350098, "learning_rate": 9.272030651340997e-05, "loss": 1.4806, "step": 288 }, { "epoch": 5.0, "grad_norm": 8.66653823852539, "learning_rate": 9.25925925925926e-05, "loss": 1.3387, "step": 290 }, { "epoch": 5.0, "eval_accuracy": 0.5757575757575758, "eval_f1_macro": 0.4401252312339328, "eval_f1_micro": 0.5757575757575758, "eval_f1_weighted": 0.5265262883858709, "eval_loss": 1.1920479536056519, "eval_precision_macro": 0.4314786059351277, "eval_precision_micro": 0.5757575757575758, "eval_precision_weighted": 0.5031131056946274, "eval_recall_macro": 0.4682644110275689, "eval_recall_micro": 0.5757575757575758, "eval_recall_weighted": 0.5757575757575758, "eval_runtime": 1.976, "eval_samples_per_second": 66.801, "eval_steps_per_second": 8.603, "step": 290 }, { "epoch": 5.0344827586206895, "grad_norm": 7.769369602203369, "learning_rate": 9.246487867177522e-05, "loss": 1.0804, "step": 292 }, { "epoch": 5.068965517241379, "grad_norm": 6.443802356719971, "learning_rate": 9.233716475095786e-05, "loss": 0.7804, "step": 294 }, { "epoch": 5.103448275862069, "grad_norm": 5.448488712310791, "learning_rate": 9.220945083014048e-05, "loss": 1.0786, "step": 296 }, { "epoch": 5.137931034482759, "grad_norm": 7.302280902862549, "learning_rate": 9.208173690932313e-05, "loss": 1.0511, "step": 298 }, { "epoch": 5.172413793103448, "grad_norm": 9.579891204833984, "learning_rate": 9.195402298850575e-05, "loss": 1.3382, "step": 300 }, { "epoch": 5.206896551724138, "grad_norm": 8.274906158447266, "learning_rate": 9.182630906768839e-05, "loss": 1.0457, "step": 302 }, { "epoch": 5.241379310344827, "grad_norm": 9.329421997070312, "learning_rate": 9.169859514687101e-05, "loss": 1.076, "step": 304 }, { "epoch": 5.275862068965517, "grad_norm": 8.050252914428711, "learning_rate": 9.157088122605364e-05, "loss": 0.744, "step": 306 }, { "epoch": 5.310344827586207, "grad_norm": 9.972197532653809, "learning_rate": 9.144316730523627e-05, "loss": 0.9465, "step": 308 }, { "epoch": 5.344827586206897, "grad_norm": 7.221776485443115, "learning_rate": 9.13154533844189e-05, "loss": 1.1189, "step": 310 }, { "epoch": 5.379310344827586, "grad_norm": 12.02575969696045, "learning_rate": 9.118773946360154e-05, "loss": 1.242, "step": 312 }, { "epoch": 5.413793103448276, "grad_norm": 10.018948554992676, "learning_rate": 9.106002554278417e-05, "loss": 0.8253, "step": 314 }, { "epoch": 5.448275862068965, "grad_norm": 16.198856353759766, "learning_rate": 9.09323116219668e-05, "loss": 2.0836, "step": 316 }, { "epoch": 5.482758620689655, "grad_norm": 15.048312187194824, "learning_rate": 9.080459770114943e-05, "loss": 1.583, "step": 318 }, { "epoch": 5.517241379310345, "grad_norm": 9.893528938293457, "learning_rate": 9.067688378033205e-05, "loss": 1.5501, "step": 320 }, { "epoch": 5.551724137931035, "grad_norm": 8.719097137451172, "learning_rate": 9.05491698595147e-05, "loss": 1.1545, "step": 322 }, { "epoch": 5.586206896551724, "grad_norm": 5.774980545043945, "learning_rate": 9.042145593869731e-05, "loss": 0.7719, "step": 324 }, { "epoch": 5.620689655172414, "grad_norm": 6.541261672973633, "learning_rate": 9.029374201787996e-05, "loss": 0.8762, "step": 326 }, { "epoch": 5.655172413793103, "grad_norm": 8.278448104858398, "learning_rate": 9.016602809706258e-05, "loss": 0.9054, "step": 328 }, { "epoch": 5.689655172413794, "grad_norm": 8.492527961730957, "learning_rate": 9.003831417624522e-05, "loss": 0.8369, "step": 330 }, { "epoch": 5.724137931034483, "grad_norm": 10.219111442565918, "learning_rate": 8.991060025542784e-05, "loss": 1.2223, "step": 332 }, { "epoch": 5.758620689655173, "grad_norm": 8.872493743896484, "learning_rate": 8.978288633461047e-05, "loss": 1.3745, "step": 334 }, { "epoch": 5.793103448275862, "grad_norm": 7.07079553604126, "learning_rate": 8.96551724137931e-05, "loss": 0.8249, "step": 336 }, { "epoch": 5.827586206896552, "grad_norm": 11.903797149658203, "learning_rate": 8.952745849297573e-05, "loss": 1.4273, "step": 338 }, { "epoch": 5.862068965517241, "grad_norm": 7.793108940124512, "learning_rate": 8.939974457215837e-05, "loss": 0.8989, "step": 340 }, { "epoch": 5.896551724137931, "grad_norm": 7.905730247497559, "learning_rate": 8.9272030651341e-05, "loss": 1.1299, "step": 342 }, { "epoch": 5.931034482758621, "grad_norm": 8.985093116760254, "learning_rate": 8.914431673052363e-05, "loss": 1.356, "step": 344 }, { "epoch": 5.9655172413793105, "grad_norm": 18.639375686645508, "learning_rate": 8.901660280970626e-05, "loss": 1.4211, "step": 346 }, { "epoch": 6.0, "grad_norm": 12.543028831481934, "learning_rate": 8.888888888888889e-05, "loss": 1.1664, "step": 348 }, { "epoch": 6.0, "eval_accuracy": 0.5075757575757576, "eval_f1_macro": 0.41789107824630456, "eval_f1_micro": 0.5075757575757576, "eval_f1_weighted": 0.49881339629823973, "eval_loss": 1.177772879600525, "eval_precision_macro": 0.506774193548387, "eval_precision_micro": 0.5075757575757576, "eval_precision_weighted": 0.5861559139784946, "eval_recall_macro": 0.43952067669172934, "eval_recall_micro": 0.5075757575757576, "eval_recall_weighted": 0.5075757575757576, "eval_runtime": 1.9845, "eval_samples_per_second": 66.515, "eval_steps_per_second": 8.566, "step": 348 }, { "epoch": 6.0344827586206895, "grad_norm": 10.878969192504883, "learning_rate": 8.876117496807152e-05, "loss": 1.2538, "step": 350 }, { "epoch": 6.068965517241379, "grad_norm": 16.26839828491211, "learning_rate": 8.863346104725416e-05, "loss": 2.0077, "step": 352 }, { "epoch": 6.103448275862069, "grad_norm": 4.248413562774658, "learning_rate": 8.850574712643679e-05, "loss": 0.6126, "step": 354 }, { "epoch": 6.137931034482759, "grad_norm": 8.243701934814453, "learning_rate": 8.837803320561942e-05, "loss": 0.9125, "step": 356 }, { "epoch": 6.172413793103448, "grad_norm": 10.138152122497559, "learning_rate": 8.825031928480205e-05, "loss": 0.7693, "step": 358 }, { "epoch": 6.206896551724138, "grad_norm": 9.491286277770996, "learning_rate": 8.812260536398468e-05, "loss": 1.1705, "step": 360 }, { "epoch": 6.241379310344827, "grad_norm": 10.823892593383789, "learning_rate": 8.79948914431673e-05, "loss": 1.4309, "step": 362 }, { "epoch": 6.275862068965517, "grad_norm": 9.899202346801758, "learning_rate": 8.786717752234995e-05, "loss": 1.0038, "step": 364 }, { "epoch": 6.310344827586207, "grad_norm": 8.473255157470703, "learning_rate": 8.773946360153256e-05, "loss": 1.1241, "step": 366 }, { "epoch": 6.344827586206897, "grad_norm": 8.142688751220703, "learning_rate": 8.761174968071521e-05, "loss": 0.9774, "step": 368 }, { "epoch": 6.379310344827586, "grad_norm": 8.222920417785645, "learning_rate": 8.748403575989783e-05, "loss": 1.105, "step": 370 }, { "epoch": 6.413793103448276, "grad_norm": 8.64330768585205, "learning_rate": 8.735632183908047e-05, "loss": 0.7421, "step": 372 }, { "epoch": 6.448275862068965, "grad_norm": 8.659346580505371, "learning_rate": 8.722860791826309e-05, "loss": 0.8458, "step": 374 }, { "epoch": 6.482758620689655, "grad_norm": 7.194849967956543, "learning_rate": 8.710089399744572e-05, "loss": 0.6802, "step": 376 }, { "epoch": 6.517241379310345, "grad_norm": 8.775270462036133, "learning_rate": 8.697318007662835e-05, "loss": 1.1627, "step": 378 }, { "epoch": 6.551724137931035, "grad_norm": 9.507420539855957, "learning_rate": 8.684546615581099e-05, "loss": 0.9029, "step": 380 }, { "epoch": 6.586206896551724, "grad_norm": 9.500436782836914, "learning_rate": 8.671775223499362e-05, "loss": 1.1927, "step": 382 }, { "epoch": 6.620689655172414, "grad_norm": 2.656595230102539, "learning_rate": 8.659003831417625e-05, "loss": 0.8011, "step": 384 }, { "epoch": 6.655172413793103, "grad_norm": 11.397368431091309, "learning_rate": 8.646232439335888e-05, "loss": 1.3364, "step": 386 }, { "epoch": 6.689655172413794, "grad_norm": 9.544795036315918, "learning_rate": 8.633461047254151e-05, "loss": 0.9104, "step": 388 }, { "epoch": 6.724137931034483, "grad_norm": 9.221898078918457, "learning_rate": 8.620689655172413e-05, "loss": 1.3124, "step": 390 }, { "epoch": 6.758620689655173, "grad_norm": 8.60506534576416, "learning_rate": 8.607918263090678e-05, "loss": 0.8603, "step": 392 }, { "epoch": 6.793103448275862, "grad_norm": 3.3715968132019043, "learning_rate": 8.59514687100894e-05, "loss": 0.8369, "step": 394 }, { "epoch": 6.827586206896552, "grad_norm": 7.430781841278076, "learning_rate": 8.582375478927204e-05, "loss": 0.7418, "step": 396 }, { "epoch": 6.862068965517241, "grad_norm": 11.155794143676758, "learning_rate": 8.569604086845466e-05, "loss": 1.2551, "step": 398 }, { "epoch": 6.896551724137931, "grad_norm": 12.489243507385254, "learning_rate": 8.55683269476373e-05, "loss": 1.3012, "step": 400 }, { "epoch": 6.931034482758621, "grad_norm": 10.465293884277344, "learning_rate": 8.544061302681992e-05, "loss": 1.122, "step": 402 }, { "epoch": 6.9655172413793105, "grad_norm": 7.462335109710693, "learning_rate": 8.531289910600255e-05, "loss": 1.2381, "step": 404 }, { "epoch": 7.0, "grad_norm": 8.508047103881836, "learning_rate": 8.518518518518518e-05, "loss": 1.1622, "step": 406 }, { "epoch": 7.0, "eval_accuracy": 0.5378787878787878, "eval_f1_macro": 0.45176980273325784, "eval_f1_micro": 0.5378787878787878, "eval_f1_weighted": 0.52507696437929, "eval_loss": 1.1722992658615112, "eval_precision_macro": 0.45142096229052747, "eval_precision_micro": 0.5378787878787878, "eval_precision_weighted": 0.5525594574507617, "eval_recall_macro": 0.4867324561403509, "eval_recall_micro": 0.5378787878787878, "eval_recall_weighted": 0.5378787878787878, "eval_runtime": 1.9976, "eval_samples_per_second": 66.081, "eval_steps_per_second": 8.51, "step": 406 }, { "epoch": 7.0344827586206895, "grad_norm": 9.81274700164795, "learning_rate": 8.505747126436782e-05, "loss": 0.8411, "step": 408 }, { "epoch": 7.068965517241379, "grad_norm": 9.398898124694824, "learning_rate": 8.492975734355045e-05, "loss": 1.0148, "step": 410 }, { "epoch": 7.103448275862069, "grad_norm": 8.494013786315918, "learning_rate": 8.480204342273308e-05, "loss": 0.5251, "step": 412 }, { "epoch": 7.137931034482759, "grad_norm": 8.39294147491455, "learning_rate": 8.467432950191571e-05, "loss": 0.6457, "step": 414 }, { "epoch": 7.172413793103448, "grad_norm": 12.938483238220215, "learning_rate": 8.454661558109834e-05, "loss": 1.0682, "step": 416 }, { "epoch": 7.206896551724138, "grad_norm": 8.100533485412598, "learning_rate": 8.441890166028097e-05, "loss": 1.0759, "step": 418 }, { "epoch": 7.241379310344827, "grad_norm": 9.130753517150879, "learning_rate": 8.42911877394636e-05, "loss": 0.665, "step": 420 }, { "epoch": 7.275862068965517, "grad_norm": 7.9572906494140625, "learning_rate": 8.416347381864624e-05, "loss": 0.5862, "step": 422 }, { "epoch": 7.310344827586207, "grad_norm": 7.2201104164123535, "learning_rate": 8.403575989782887e-05, "loss": 0.8631, "step": 424 }, { "epoch": 7.344827586206897, "grad_norm": 9.218771934509277, "learning_rate": 8.39080459770115e-05, "loss": 0.9256, "step": 426 }, { "epoch": 7.379310344827586, "grad_norm": 10.637073516845703, "learning_rate": 8.378033205619413e-05, "loss": 0.5491, "step": 428 }, { "epoch": 7.413793103448276, "grad_norm": 6.346437931060791, "learning_rate": 8.365261813537676e-05, "loss": 0.7555, "step": 430 }, { "epoch": 7.448275862068965, "grad_norm": 5.5799384117126465, "learning_rate": 8.35249042145594e-05, "loss": 0.3958, "step": 432 }, { "epoch": 7.482758620689655, "grad_norm": 11.625855445861816, "learning_rate": 8.339719029374203e-05, "loss": 0.7716, "step": 434 }, { "epoch": 7.517241379310345, "grad_norm": 11.031689643859863, "learning_rate": 8.326947637292465e-05, "loss": 0.946, "step": 436 }, { "epoch": 7.551724137931035, "grad_norm": 14.949385643005371, "learning_rate": 8.314176245210729e-05, "loss": 1.448, "step": 438 }, { "epoch": 7.586206896551724, "grad_norm": 9.996256828308105, "learning_rate": 8.301404853128991e-05, "loss": 1.0621, "step": 440 }, { "epoch": 7.620689655172414, "grad_norm": 12.930780410766602, "learning_rate": 8.288633461047255e-05, "loss": 1.5245, "step": 442 }, { "epoch": 7.655172413793103, "grad_norm": 9.777554512023926, "learning_rate": 8.275862068965517e-05, "loss": 0.7166, "step": 444 }, { "epoch": 7.689655172413794, "grad_norm": 15.452765464782715, "learning_rate": 8.263090676883782e-05, "loss": 0.81, "step": 446 }, { "epoch": 7.724137931034483, "grad_norm": 5.021923542022705, "learning_rate": 8.250319284802044e-05, "loss": 0.4846, "step": 448 }, { "epoch": 7.758620689655173, "grad_norm": 10.42216682434082, "learning_rate": 8.237547892720307e-05, "loss": 0.8314, "step": 450 }, { "epoch": 7.793103448275862, "grad_norm": 10.091817855834961, "learning_rate": 8.22477650063857e-05, "loss": 0.7462, "step": 452 }, { "epoch": 7.827586206896552, "grad_norm": 10.696878433227539, "learning_rate": 8.212005108556833e-05, "loss": 1.1465, "step": 454 }, { "epoch": 7.862068965517241, "grad_norm": 13.578857421875, "learning_rate": 8.199233716475096e-05, "loss": 1.7135, "step": 456 }, { "epoch": 7.896551724137931, "grad_norm": 8.861135482788086, "learning_rate": 8.18646232439336e-05, "loss": 0.9397, "step": 458 }, { "epoch": 7.931034482758621, "grad_norm": 8.846278190612793, "learning_rate": 8.173690932311623e-05, "loss": 0.932, "step": 460 }, { "epoch": 7.9655172413793105, "grad_norm": 13.574858665466309, "learning_rate": 8.160919540229886e-05, "loss": 1.6147, "step": 462 }, { "epoch": 8.0, "grad_norm": 11.378049850463867, "learning_rate": 8.148148148148148e-05, "loss": 0.9827, "step": 464 }, { "epoch": 8.0, "eval_accuracy": 0.6212121212121212, "eval_f1_macro": 0.5083652151343768, "eval_f1_micro": 0.6212121212121212, "eval_f1_weighted": 0.6074328355724158, "eval_loss": 1.0619021654129028, "eval_precision_macro": 0.5036885527157723, "eval_precision_micro": 0.6212121212121212, "eval_precision_weighted": 0.61396284194398, "eval_recall_macro": 0.5344924812030075, "eval_recall_micro": 0.6212121212121212, "eval_recall_weighted": 0.6212121212121212, "eval_runtime": 1.9917, "eval_samples_per_second": 66.275, "eval_steps_per_second": 8.535, "step": 464 }, { "epoch": 8.03448275862069, "grad_norm": 5.481504440307617, "learning_rate": 8.135376756066412e-05, "loss": 0.4917, "step": 466 }, { "epoch": 8.068965517241379, "grad_norm": 7.311232566833496, "learning_rate": 8.122605363984674e-05, "loss": 0.7269, "step": 468 }, { "epoch": 8.10344827586207, "grad_norm": 9.246880531311035, "learning_rate": 8.109833971902938e-05, "loss": 0.8737, "step": 470 }, { "epoch": 8.137931034482758, "grad_norm": 8.400725364685059, "learning_rate": 8.0970625798212e-05, "loss": 0.5171, "step": 472 }, { "epoch": 8.172413793103448, "grad_norm": 7.584453105926514, "learning_rate": 8.084291187739465e-05, "loss": 0.4877, "step": 474 }, { "epoch": 8.206896551724139, "grad_norm": 3.6026341915130615, "learning_rate": 8.071519795657727e-05, "loss": 0.3995, "step": 476 }, { "epoch": 8.241379310344827, "grad_norm": 11.562987327575684, "learning_rate": 8.05874840357599e-05, "loss": 0.7385, "step": 478 }, { "epoch": 8.275862068965518, "grad_norm": 6.191165447235107, "learning_rate": 8.045977011494253e-05, "loss": 0.5029, "step": 480 }, { "epoch": 8.310344827586206, "grad_norm": 6.5581583976745605, "learning_rate": 8.033205619412516e-05, "loss": 0.4527, "step": 482 }, { "epoch": 8.344827586206897, "grad_norm": 5.58575963973999, "learning_rate": 8.020434227330779e-05, "loss": 0.505, "step": 484 }, { "epoch": 8.379310344827585, "grad_norm": 7.359447956085205, "learning_rate": 8.007662835249042e-05, "loss": 0.3806, "step": 486 }, { "epoch": 8.413793103448276, "grad_norm": 9.096012115478516, "learning_rate": 7.994891443167306e-05, "loss": 0.7836, "step": 488 }, { "epoch": 8.448275862068966, "grad_norm": 5.758499622344971, "learning_rate": 7.982120051085569e-05, "loss": 0.6284, "step": 490 }, { "epoch": 8.482758620689655, "grad_norm": 9.046234130859375, "learning_rate": 7.969348659003832e-05, "loss": 0.8215, "step": 492 }, { "epoch": 8.517241379310345, "grad_norm": 10.501338958740234, "learning_rate": 7.956577266922095e-05, "loss": 0.7549, "step": 494 }, { "epoch": 8.551724137931034, "grad_norm": 12.925354957580566, "learning_rate": 7.943805874840358e-05, "loss": 0.9301, "step": 496 }, { "epoch": 8.586206896551724, "grad_norm": 9.080878257751465, "learning_rate": 7.931034482758621e-05, "loss": 0.9217, "step": 498 }, { "epoch": 8.620689655172415, "grad_norm": 11.387174606323242, "learning_rate": 7.918263090676885e-05, "loss": 1.1331, "step": 500 }, { "epoch": 8.655172413793103, "grad_norm": 4.864457130432129, "learning_rate": 7.905491698595148e-05, "loss": 0.3399, "step": 502 }, { "epoch": 8.689655172413794, "grad_norm": 9.243449211120605, "learning_rate": 7.892720306513411e-05, "loss": 0.8055, "step": 504 }, { "epoch": 8.724137931034482, "grad_norm": 13.595425605773926, "learning_rate": 7.879948914431673e-05, "loss": 1.2361, "step": 506 }, { "epoch": 8.758620689655173, "grad_norm": 9.03889274597168, "learning_rate": 7.867177522349937e-05, "loss": 1.2648, "step": 508 }, { "epoch": 8.793103448275861, "grad_norm": 10.691272735595703, "learning_rate": 7.854406130268199e-05, "loss": 0.8417, "step": 510 }, { "epoch": 8.827586206896552, "grad_norm": 15.45632266998291, "learning_rate": 7.841634738186464e-05, "loss": 1.1186, "step": 512 }, { "epoch": 8.862068965517242, "grad_norm": 10.35450267791748, "learning_rate": 7.828863346104725e-05, "loss": 1.0734, "step": 514 }, { "epoch": 8.89655172413793, "grad_norm": 5.785059452056885, "learning_rate": 7.81609195402299e-05, "loss": 0.5601, "step": 516 }, { "epoch": 8.931034482758621, "grad_norm": 8.064878463745117, "learning_rate": 7.803320561941252e-05, "loss": 0.8273, "step": 518 }, { "epoch": 8.96551724137931, "grad_norm": 9.697986602783203, "learning_rate": 7.790549169859515e-05, "loss": 0.7774, "step": 520 }, { "epoch": 9.0, "grad_norm": 14.628928184509277, "learning_rate": 7.777777777777778e-05, "loss": 1.3416, "step": 522 }, { "epoch": 9.0, "eval_accuracy": 0.5, "eval_f1_macro": 0.3996626928502454, "eval_f1_micro": 0.5, "eval_f1_weighted": 0.4689747329928523, "eval_loss": 1.3995325565338135, "eval_precision_macro": 0.4217794875689612, "eval_precision_micro": 0.5, "eval_precision_weighted": 0.5023755373635757, "eval_recall_macro": 0.45092418546365914, "eval_recall_micro": 0.5, "eval_recall_weighted": 0.5, "eval_runtime": 1.9802, "eval_samples_per_second": 66.659, "eval_steps_per_second": 8.585, "step": 522 }, { "epoch": 9.03448275862069, "grad_norm": 8.5454683303833, "learning_rate": 7.765006385696041e-05, "loss": 0.7221, "step": 524 }, { "epoch": 9.068965517241379, "grad_norm": 12.215627670288086, "learning_rate": 7.752234993614304e-05, "loss": 1.1389, "step": 526 }, { "epoch": 9.10344827586207, "grad_norm": 7.772884368896484, "learning_rate": 7.739463601532568e-05, "loss": 0.5547, "step": 528 }, { "epoch": 9.137931034482758, "grad_norm": 6.657981872558594, "learning_rate": 7.726692209450831e-05, "loss": 0.4854, "step": 530 }, { "epoch": 9.172413793103448, "grad_norm": 4.549035549163818, "learning_rate": 7.713920817369094e-05, "loss": 0.4071, "step": 532 }, { "epoch": 9.206896551724139, "grad_norm": 10.139058113098145, "learning_rate": 7.701149425287356e-05, "loss": 0.8507, "step": 534 }, { "epoch": 9.241379310344827, "grad_norm": 9.346701622009277, "learning_rate": 7.68837803320562e-05, "loss": 0.5146, "step": 536 }, { "epoch": 9.275862068965518, "grad_norm": 8.409459114074707, "learning_rate": 7.675606641123882e-05, "loss": 0.7852, "step": 538 }, { "epoch": 9.310344827586206, "grad_norm": 10.624256134033203, "learning_rate": 7.662835249042147e-05, "loss": 0.7101, "step": 540 }, { "epoch": 9.344827586206897, "grad_norm": 10.048233032226562, "learning_rate": 7.650063856960408e-05, "loss": 0.9228, "step": 542 }, { "epoch": 9.379310344827585, "grad_norm": 7.511538982391357, "learning_rate": 7.637292464878673e-05, "loss": 0.4572, "step": 544 }, { "epoch": 9.413793103448276, "grad_norm": 5.808769226074219, "learning_rate": 7.624521072796935e-05, "loss": 0.3693, "step": 546 }, { "epoch": 9.448275862068966, "grad_norm": 7.7615156173706055, "learning_rate": 7.611749680715198e-05, "loss": 0.5247, "step": 548 }, { "epoch": 9.482758620689655, "grad_norm": 13.348309516906738, "learning_rate": 7.598978288633461e-05, "loss": 1.0443, "step": 550 }, { "epoch": 9.517241379310345, "grad_norm": 9.64974594116211, "learning_rate": 7.586206896551724e-05, "loss": 0.563, "step": 552 }, { "epoch": 9.551724137931034, "grad_norm": 11.617941856384277, "learning_rate": 7.573435504469987e-05, "loss": 1.021, "step": 554 }, { "epoch": 9.586206896551724, "grad_norm": 8.760254859924316, "learning_rate": 7.56066411238825e-05, "loss": 0.7207, "step": 556 }, { "epoch": 9.620689655172415, "grad_norm": 9.84865951538086, "learning_rate": 7.547892720306514e-05, "loss": 0.4963, "step": 558 }, { "epoch": 9.655172413793103, "grad_norm": 9.160222053527832, "learning_rate": 7.535121328224777e-05, "loss": 0.8536, "step": 560 }, { "epoch": 9.689655172413794, "grad_norm": 6.880321502685547, "learning_rate": 7.522349936143039e-05, "loss": 1.1081, "step": 562 }, { "epoch": 9.724137931034482, "grad_norm": 7.537998676300049, "learning_rate": 7.509578544061303e-05, "loss": 0.5926, "step": 564 }, { "epoch": 9.758620689655173, "grad_norm": 10.582006454467773, "learning_rate": 7.496807151979565e-05, "loss": 0.9216, "step": 566 }, { "epoch": 9.793103448275861, "grad_norm": 8.612391471862793, "learning_rate": 7.48403575989783e-05, "loss": 0.4138, "step": 568 }, { "epoch": 9.827586206896552, "grad_norm": 9.592196464538574, "learning_rate": 7.471264367816091e-05, "loss": 0.7652, "step": 570 }, { "epoch": 9.862068965517242, "grad_norm": 7.557244300842285, "learning_rate": 7.458492975734356e-05, "loss": 0.4475, "step": 572 }, { "epoch": 9.89655172413793, "grad_norm": 12.449685096740723, "learning_rate": 7.445721583652618e-05, "loss": 0.7721, "step": 574 }, { "epoch": 9.931034482758621, "grad_norm": 7.203516960144043, "learning_rate": 7.432950191570882e-05, "loss": 1.0207, "step": 576 }, { "epoch": 9.96551724137931, "grad_norm": 4.533941745758057, "learning_rate": 7.420178799489144e-05, "loss": 0.5956, "step": 578 }, { "epoch": 10.0, "grad_norm": 21.596515655517578, "learning_rate": 7.407407407407407e-05, "loss": 0.758, "step": 580 }, { "epoch": 10.0, "eval_accuracy": 0.5984848484848485, "eval_f1_macro": 0.5065544106205532, "eval_f1_micro": 0.5984848484848485, "eval_f1_weighted": 0.5835752856194464, "eval_loss": 1.1693105697631836, "eval_precision_macro": 0.5261672850958565, "eval_precision_micro": 0.5984848484848485, "eval_precision_weighted": 0.6030692476715204, "eval_recall_macro": 0.5278665413533835, "eval_recall_micro": 0.5984848484848485, "eval_recall_weighted": 0.5984848484848485, "eval_runtime": 2.0002, "eval_samples_per_second": 65.992, "eval_steps_per_second": 8.499, "step": 580 }, { "epoch": 10.03448275862069, "grad_norm": 5.144809246063232, "learning_rate": 7.39463601532567e-05, "loss": 0.2258, "step": 582 }, { "epoch": 10.068965517241379, "grad_norm": 8.242935180664062, "learning_rate": 7.381864623243934e-05, "loss": 0.45, "step": 584 }, { "epoch": 10.10344827586207, "grad_norm": 3.4672491550445557, "learning_rate": 7.369093231162197e-05, "loss": 0.4987, "step": 586 }, { "epoch": 10.137931034482758, "grad_norm": 1.9440521001815796, "learning_rate": 7.35632183908046e-05, "loss": 0.2707, "step": 588 }, { "epoch": 10.172413793103448, "grad_norm": 11.417020797729492, "learning_rate": 7.343550446998723e-05, "loss": 0.6887, "step": 590 }, { "epoch": 10.206896551724139, "grad_norm": 6.448490142822266, "learning_rate": 7.330779054916986e-05, "loss": 0.3351, "step": 592 }, { "epoch": 10.241379310344827, "grad_norm": 21.801414489746094, "learning_rate": 7.31800766283525e-05, "loss": 0.8969, "step": 594 }, { "epoch": 10.275862068965518, "grad_norm": 3.7770888805389404, "learning_rate": 7.305236270753513e-05, "loss": 0.3175, "step": 596 }, { "epoch": 10.310344827586206, "grad_norm": 12.79362678527832, "learning_rate": 7.292464878671776e-05, "loss": 0.756, "step": 598 }, { "epoch": 10.344827586206897, "grad_norm": 12.172070503234863, "learning_rate": 7.279693486590039e-05, "loss": 0.6943, "step": 600 }, { "epoch": 10.379310344827585, "grad_norm": 11.363933563232422, "learning_rate": 7.266922094508302e-05, "loss": 0.7048, "step": 602 }, { "epoch": 10.413793103448276, "grad_norm": 7.672904968261719, "learning_rate": 7.254150702426565e-05, "loss": 0.2959, "step": 604 }, { "epoch": 10.448275862068966, "grad_norm": 9.661025047302246, "learning_rate": 7.241379310344828e-05, "loss": 0.4829, "step": 606 }, { "epoch": 10.482758620689655, "grad_norm": 8.507035255432129, "learning_rate": 7.22860791826309e-05, "loss": 0.2263, "step": 608 }, { "epoch": 10.517241379310345, "grad_norm": 6.187190055847168, "learning_rate": 7.215836526181355e-05, "loss": 0.5414, "step": 610 }, { "epoch": 10.551724137931034, "grad_norm": 11.629401206970215, "learning_rate": 7.203065134099617e-05, "loss": 0.7474, "step": 612 }, { "epoch": 10.586206896551724, "grad_norm": 7.820981025695801, "learning_rate": 7.190293742017881e-05, "loss": 0.4997, "step": 614 }, { "epoch": 10.620689655172415, "grad_norm": 14.111104011535645, "learning_rate": 7.177522349936143e-05, "loss": 0.4296, "step": 616 }, { "epoch": 10.655172413793103, "grad_norm": 10.963750839233398, "learning_rate": 7.164750957854408e-05, "loss": 0.4871, "step": 618 }, { "epoch": 10.689655172413794, "grad_norm": 17.322771072387695, "learning_rate": 7.151979565772669e-05, "loss": 0.9553, "step": 620 }, { "epoch": 10.724137931034482, "grad_norm": 9.986748695373535, "learning_rate": 7.139208173690932e-05, "loss": 0.5509, "step": 622 }, { "epoch": 10.758620689655173, "grad_norm": 9.360352516174316, "learning_rate": 7.126436781609196e-05, "loss": 0.6703, "step": 624 }, { "epoch": 10.793103448275861, "grad_norm": 3.3035664558410645, "learning_rate": 7.113665389527459e-05, "loss": 0.4547, "step": 626 }, { "epoch": 10.827586206896552, "grad_norm": 8.55704116821289, "learning_rate": 7.100893997445722e-05, "loss": 0.7755, "step": 628 }, { "epoch": 10.862068965517242, "grad_norm": 12.647590637207031, "learning_rate": 7.088122605363985e-05, "loss": 0.8421, "step": 630 }, { "epoch": 10.89655172413793, "grad_norm": 4.907244682312012, "learning_rate": 7.075351213282248e-05, "loss": 0.3781, "step": 632 }, { "epoch": 10.931034482758621, "grad_norm": 10.101003646850586, "learning_rate": 7.062579821200511e-05, "loss": 0.6021, "step": 634 }, { "epoch": 10.96551724137931, "grad_norm": 15.203731536865234, "learning_rate": 7.049808429118773e-05, "loss": 0.9916, "step": 636 }, { "epoch": 11.0, "grad_norm": 20.572900772094727, "learning_rate": 7.037037037037038e-05, "loss": 0.7758, "step": 638 }, { "epoch": 11.0, "eval_accuracy": 0.6515151515151515, "eval_f1_macro": 0.5490749541247198, "eval_f1_micro": 0.6515151515151515, "eval_f1_weighted": 0.6319976363055722, "eval_loss": 1.0799717903137207, "eval_precision_macro": 0.5729282057650646, "eval_precision_micro": 0.6515151515151515, "eval_precision_weighted": 0.6500887606871381, "eval_recall_macro": 0.5710213032581454, "eval_recall_micro": 0.6515151515151515, "eval_recall_weighted": 0.6515151515151515, "eval_runtime": 1.9879, "eval_samples_per_second": 66.401, "eval_steps_per_second": 8.552, "step": 638 }, { "epoch": 11.03448275862069, "grad_norm": 6.214511871337891, "learning_rate": 7.0242656449553e-05, "loss": 0.4373, "step": 640 }, { "epoch": 11.068965517241379, "grad_norm": 13.804490089416504, "learning_rate": 7.011494252873564e-05, "loss": 0.7242, "step": 642 }, { "epoch": 11.10344827586207, "grad_norm": 7.258790969848633, "learning_rate": 6.998722860791826e-05, "loss": 0.3412, "step": 644 }, { "epoch": 11.137931034482758, "grad_norm": 13.48838996887207, "learning_rate": 6.98595146871009e-05, "loss": 0.7093, "step": 646 }, { "epoch": 11.172413793103448, "grad_norm": 3.5530500411987305, "learning_rate": 6.973180076628352e-05, "loss": 0.1557, "step": 648 }, { "epoch": 11.206896551724139, "grad_norm": 5.173003196716309, "learning_rate": 6.960408684546615e-05, "loss": 0.2043, "step": 650 }, { "epoch": 11.241379310344827, "grad_norm": 8.948324203491211, "learning_rate": 6.947637292464879e-05, "loss": 0.3718, "step": 652 }, { "epoch": 11.275862068965518, "grad_norm": 13.394603729248047, "learning_rate": 6.934865900383142e-05, "loss": 0.4858, "step": 654 }, { "epoch": 11.310344827586206, "grad_norm": 3.575120210647583, "learning_rate": 6.922094508301405e-05, "loss": 0.2133, "step": 656 }, { "epoch": 11.344827586206897, "grad_norm": 6.710623264312744, "learning_rate": 6.909323116219668e-05, "loss": 0.4072, "step": 658 }, { "epoch": 11.379310344827585, "grad_norm": 14.04590892791748, "learning_rate": 6.896551724137931e-05, "loss": 0.9938, "step": 660 }, { "epoch": 11.413793103448276, "grad_norm": 7.6388444900512695, "learning_rate": 6.883780332056194e-05, "loss": 0.4864, "step": 662 }, { "epoch": 11.448275862068966, "grad_norm": 6.261629104614258, "learning_rate": 6.871008939974458e-05, "loss": 0.3112, "step": 664 }, { "epoch": 11.482758620689655, "grad_norm": 7.990070819854736, "learning_rate": 6.858237547892721e-05, "loss": 0.6052, "step": 666 }, { "epoch": 11.517241379310345, "grad_norm": 5.1240949630737305, "learning_rate": 6.845466155810984e-05, "loss": 0.6029, "step": 668 }, { "epoch": 11.551724137931034, "grad_norm": 3.3157217502593994, "learning_rate": 6.832694763729247e-05, "loss": 0.2027, "step": 670 }, { "epoch": 11.586206896551724, "grad_norm": 9.071480751037598, "learning_rate": 6.81992337164751e-05, "loss": 0.257, "step": 672 }, { "epoch": 11.620689655172415, "grad_norm": 13.492743492126465, "learning_rate": 6.807151979565773e-05, "loss": 0.4914, "step": 674 }, { "epoch": 11.655172413793103, "grad_norm": 1.5881291627883911, "learning_rate": 6.794380587484037e-05, "loss": 0.3057, "step": 676 }, { "epoch": 11.689655172413794, "grad_norm": 13.12863540649414, "learning_rate": 6.781609195402298e-05, "loss": 0.4456, "step": 678 }, { "epoch": 11.724137931034482, "grad_norm": 15.156206130981445, "learning_rate": 6.768837803320563e-05, "loss": 0.4095, "step": 680 }, { "epoch": 11.758620689655173, "grad_norm": 12.477018356323242, "learning_rate": 6.756066411238825e-05, "loss": 0.6728, "step": 682 }, { "epoch": 11.793103448275861, "grad_norm": 6.536929607391357, "learning_rate": 6.74329501915709e-05, "loss": 0.3572, "step": 684 }, { "epoch": 11.827586206896552, "grad_norm": 8.295530319213867, "learning_rate": 6.730523627075351e-05, "loss": 0.3506, "step": 686 }, { "epoch": 11.862068965517242, "grad_norm": 2.183727979660034, "learning_rate": 6.717752234993616e-05, "loss": 0.3145, "step": 688 }, { "epoch": 11.89655172413793, "grad_norm": 9.358091354370117, "learning_rate": 6.704980842911877e-05, "loss": 0.5877, "step": 690 }, { "epoch": 11.931034482758621, "grad_norm": 7.218227386474609, "learning_rate": 6.69220945083014e-05, "loss": 0.4683, "step": 692 }, { "epoch": 11.96551724137931, "grad_norm": 6.784433364868164, "learning_rate": 6.679438058748404e-05, "loss": 0.3328, "step": 694 }, { "epoch": 12.0, "grad_norm": 4.573869228363037, "learning_rate": 6.666666666666667e-05, "loss": 0.2319, "step": 696 }, { "epoch": 12.0, "eval_accuracy": 0.6742424242424242, "eval_f1_macro": 0.546672932330827, "eval_f1_micro": 0.6742424242424242, "eval_f1_weighted": 0.6409964779372674, "eval_loss": 1.1553009748458862, "eval_precision_macro": 0.5816405209018068, "eval_precision_micro": 0.6742424242424242, "eval_precision_weighted": 0.6699464486928733, "eval_recall_macro": 0.5711466165413533, "eval_recall_micro": 0.6742424242424242, "eval_recall_weighted": 0.6742424242424242, "eval_runtime": 1.9872, "eval_samples_per_second": 66.426, "eval_steps_per_second": 8.555, "step": 696 }, { "epoch": 12.03448275862069, "grad_norm": 1.5767076015472412, "learning_rate": 6.65389527458493e-05, "loss": 0.235, "step": 698 }, { "epoch": 12.068965517241379, "grad_norm": 7.686869144439697, "learning_rate": 6.641123882503193e-05, "loss": 0.5692, "step": 700 }, { "epoch": 12.10344827586207, "grad_norm": 13.89122200012207, "learning_rate": 6.628352490421456e-05, "loss": 0.4877, "step": 702 }, { "epoch": 12.137931034482758, "grad_norm": 4.811250686645508, "learning_rate": 6.61558109833972e-05, "loss": 0.2121, "step": 704 }, { "epoch": 12.172413793103448, "grad_norm": 2.0311198234558105, "learning_rate": 6.602809706257981e-05, "loss": 0.2037, "step": 706 }, { "epoch": 12.206896551724139, "grad_norm": 7.1766133308410645, "learning_rate": 6.590038314176246e-05, "loss": 0.2169, "step": 708 }, { "epoch": 12.241379310344827, "grad_norm": 5.523502826690674, "learning_rate": 6.577266922094508e-05, "loss": 1.335, "step": 710 }, { "epoch": 12.275862068965518, "grad_norm": 16.38228416442871, "learning_rate": 6.564495530012772e-05, "loss": 0.7502, "step": 712 }, { "epoch": 12.310344827586206, "grad_norm": 8.219087600708008, "learning_rate": 6.551724137931034e-05, "loss": 0.7137, "step": 714 }, { "epoch": 12.344827586206897, "grad_norm": 4.552286148071289, "learning_rate": 6.538952745849299e-05, "loss": 0.2233, "step": 716 }, { "epoch": 12.379310344827585, "grad_norm": 2.2111048698425293, "learning_rate": 6.52618135376756e-05, "loss": 0.1856, "step": 718 }, { "epoch": 12.413793103448276, "grad_norm": 10.02652359008789, "learning_rate": 6.513409961685824e-05, "loss": 0.4808, "step": 720 }, { "epoch": 12.448275862068966, "grad_norm": 4.486725807189941, "learning_rate": 6.500638569604087e-05, "loss": 0.4057, "step": 722 }, { "epoch": 12.482758620689655, "grad_norm": 1.8516465425491333, "learning_rate": 6.48786717752235e-05, "loss": 0.1974, "step": 724 }, { "epoch": 12.517241379310345, "grad_norm": 6.200466156005859, "learning_rate": 6.475095785440613e-05, "loss": 0.2568, "step": 726 }, { "epoch": 12.551724137931034, "grad_norm": 16.36850929260254, "learning_rate": 6.462324393358876e-05, "loss": 0.5874, "step": 728 }, { "epoch": 12.586206896551724, "grad_norm": 4.7899346351623535, "learning_rate": 6.44955300127714e-05, "loss": 0.3983, "step": 730 }, { "epoch": 12.620689655172415, "grad_norm": 7.357030868530273, "learning_rate": 6.436781609195403e-05, "loss": 0.3513, "step": 732 }, { "epoch": 12.655172413793103, "grad_norm": 5.966126441955566, "learning_rate": 6.424010217113666e-05, "loss": 0.4016, "step": 734 }, { "epoch": 12.689655172413794, "grad_norm": 6.382541656494141, "learning_rate": 6.411238825031929e-05, "loss": 0.1682, "step": 736 }, { "epoch": 12.724137931034482, "grad_norm": 8.352551460266113, "learning_rate": 6.398467432950191e-05, "loss": 0.3286, "step": 738 }, { "epoch": 12.758620689655173, "grad_norm": 9.189525604248047, "learning_rate": 6.385696040868455e-05, "loss": 0.49, "step": 740 }, { "epoch": 12.793103448275861, "grad_norm": 15.152857780456543, "learning_rate": 6.372924648786717e-05, "loss": 0.1309, "step": 742 }, { "epoch": 12.827586206896552, "grad_norm": 3.6843764781951904, "learning_rate": 6.360153256704982e-05, "loss": 0.1974, "step": 744 }, { "epoch": 12.862068965517242, "grad_norm": 10.768837928771973, "learning_rate": 6.347381864623243e-05, "loss": 0.8215, "step": 746 }, { "epoch": 12.89655172413793, "grad_norm": 7.068192005157471, "learning_rate": 6.334610472541508e-05, "loss": 0.2304, "step": 748 }, { "epoch": 12.931034482758621, "grad_norm": 12.653234481811523, "learning_rate": 6.32183908045977e-05, "loss": 0.5768, "step": 750 }, { "epoch": 12.96551724137931, "grad_norm": 7.09431266784668, "learning_rate": 6.309067688378033e-05, "loss": 0.6238, "step": 752 }, { "epoch": 13.0, "grad_norm": 7.309642791748047, "learning_rate": 6.296296296296296e-05, "loss": 0.3528, "step": 754 }, { "epoch": 13.0, "eval_accuracy": 0.6893939393939394, "eval_f1_macro": 0.5794293474965745, "eval_f1_micro": 0.6893939393939394, "eval_f1_weighted": 0.6711440544996694, "eval_loss": 1.1685433387756348, "eval_precision_macro": 0.5887005339770317, "eval_precision_micro": 0.6893939393939394, "eval_precision_weighted": 0.6752094679514034, "eval_recall_macro": 0.5955043859649124, "eval_recall_micro": 0.6893939393939394, "eval_recall_weighted": 0.6893939393939394, "eval_runtime": 1.9392, "eval_samples_per_second": 68.07, "eval_steps_per_second": 8.767, "step": 754 }, { "epoch": 13.03448275862069, "grad_norm": 9.388544082641602, "learning_rate": 6.283524904214559e-05, "loss": 0.6725, "step": 756 }, { "epoch": 13.068965517241379, "grad_norm": 15.097204208374023, "learning_rate": 6.270753512132822e-05, "loss": 0.6723, "step": 758 }, { "epoch": 13.10344827586207, "grad_norm": 7.6261796951293945, "learning_rate": 6.257982120051086e-05, "loss": 0.3793, "step": 760 }, { "epoch": 13.137931034482758, "grad_norm": 2.728222370147705, "learning_rate": 6.245210727969349e-05, "loss": 0.1291, "step": 762 }, { "epoch": 13.172413793103448, "grad_norm": 7.391420841217041, "learning_rate": 6.232439335887612e-05, "loss": 0.4385, "step": 764 }, { "epoch": 13.206896551724139, "grad_norm": 9.982900619506836, "learning_rate": 6.219667943805875e-05, "loss": 0.3353, "step": 766 }, { "epoch": 13.241379310344827, "grad_norm": 7.945486545562744, "learning_rate": 6.206896551724138e-05, "loss": 0.1623, "step": 768 }, { "epoch": 13.275862068965518, "grad_norm": 11.18921184539795, "learning_rate": 6.194125159642401e-05, "loss": 0.1834, "step": 770 }, { "epoch": 13.310344827586206, "grad_norm": 1.6263447999954224, "learning_rate": 6.181353767560665e-05, "loss": 0.0572, "step": 772 }, { "epoch": 13.344827586206897, "grad_norm": 8.16580581665039, "learning_rate": 6.168582375478928e-05, "loss": 0.3301, "step": 774 }, { "epoch": 13.379310344827585, "grad_norm": 2.9834048748016357, "learning_rate": 6.155810983397191e-05, "loss": 0.1367, "step": 776 }, { "epoch": 13.413793103448276, "grad_norm": 9.801024436950684, "learning_rate": 6.143039591315454e-05, "loss": 0.2472, "step": 778 }, { "epoch": 13.448275862068966, "grad_norm": 13.256599426269531, "learning_rate": 6.130268199233716e-05, "loss": 0.6242, "step": 780 }, { "epoch": 13.482758620689655, "grad_norm": 12.600865364074707, "learning_rate": 6.11749680715198e-05, "loss": 0.387, "step": 782 }, { "epoch": 13.517241379310345, "grad_norm": 12.394946098327637, "learning_rate": 6.104725415070242e-05, "loss": 0.3973, "step": 784 }, { "epoch": 13.551724137931034, "grad_norm": 16.835023880004883, "learning_rate": 6.091954022988506e-05, "loss": 0.5736, "step": 786 }, { "epoch": 13.586206896551724, "grad_norm": 10.348804473876953, "learning_rate": 6.0791826309067686e-05, "loss": 0.3203, "step": 788 }, { "epoch": 13.620689655172415, "grad_norm": 10.827553749084473, "learning_rate": 6.0664112388250325e-05, "loss": 0.3311, "step": 790 }, { "epoch": 13.655172413793103, "grad_norm": 4.213782787322998, "learning_rate": 6.053639846743295e-05, "loss": 0.0665, "step": 792 }, { "epoch": 13.689655172413794, "grad_norm": 5.154265403747559, "learning_rate": 6.040868454661558e-05, "loss": 0.3031, "step": 794 }, { "epoch": 13.724137931034482, "grad_norm": 18.196022033691406, "learning_rate": 6.028097062579821e-05, "loss": 0.8001, "step": 796 }, { "epoch": 13.758620689655173, "grad_norm": 3.6230857372283936, "learning_rate": 6.0153256704980845e-05, "loss": 0.1406, "step": 798 }, { "epoch": 13.793103448275861, "grad_norm": 0.9606171250343323, "learning_rate": 6.0025542784163477e-05, "loss": 0.0544, "step": 800 }, { "epoch": 13.827586206896552, "grad_norm": 13.326411247253418, "learning_rate": 5.989782886334611e-05, "loss": 0.2442, "step": 802 }, { "epoch": 13.862068965517242, "grad_norm": 8.518882751464844, "learning_rate": 5.977011494252874e-05, "loss": 0.2051, "step": 804 }, { "epoch": 13.89655172413793, "grad_norm": 4.24729585647583, "learning_rate": 5.964240102171137e-05, "loss": 0.4868, "step": 806 }, { "epoch": 13.931034482758621, "grad_norm": 0.7721981406211853, "learning_rate": 5.9514687100893996e-05, "loss": 0.1428, "step": 808 }, { "epoch": 13.96551724137931, "grad_norm": 11.86274242401123, "learning_rate": 5.9386973180076635e-05, "loss": 0.7225, "step": 810 }, { "epoch": 14.0, "grad_norm": 14.592938423156738, "learning_rate": 5.925925925925926e-05, "loss": 0.6238, "step": 812 }, { "epoch": 14.0, "eval_accuracy": 0.6439393939393939, "eval_f1_macro": 0.5579426867483074, "eval_f1_micro": 0.6439393939393939, "eval_f1_weighted": 0.6285409584589912, "eval_loss": 1.1781333684921265, "eval_precision_macro": 0.5450511546824911, "eval_precision_micro": 0.6439393939393939, "eval_precision_weighted": 0.6277759353712432, "eval_recall_macro": 0.5856046365914788, "eval_recall_micro": 0.6439393939393939, "eval_recall_weighted": 0.6439393939393939, "eval_runtime": 1.9417, "eval_samples_per_second": 67.982, "eval_steps_per_second": 8.755, "step": 812 }, { "epoch": 14.03448275862069, "grad_norm": 11.65465259552002, "learning_rate": 5.91315453384419e-05, "loss": 0.3351, "step": 814 }, { "epoch": 14.068965517241379, "grad_norm": 9.413543701171875, "learning_rate": 5.900383141762452e-05, "loss": 0.3187, "step": 816 }, { "epoch": 14.10344827586207, "grad_norm": 3.177682638168335, "learning_rate": 5.887611749680716e-05, "loss": 0.0781, "step": 818 }, { "epoch": 14.137931034482758, "grad_norm": 9.098755836486816, "learning_rate": 5.8748403575989787e-05, "loss": 0.1169, "step": 820 }, { "epoch": 14.172413793103448, "grad_norm": 1.7022143602371216, "learning_rate": 5.862068965517241e-05, "loss": 0.2369, "step": 822 }, { "epoch": 14.206896551724139, "grad_norm": 6.126805782318115, "learning_rate": 5.849297573435505e-05, "loss": 0.3991, "step": 824 }, { "epoch": 14.241379310344827, "grad_norm": 6.332644939422607, "learning_rate": 5.8365261813537675e-05, "loss": 0.1479, "step": 826 }, { "epoch": 14.275862068965518, "grad_norm": 5.615021228790283, "learning_rate": 5.823754789272031e-05, "loss": 0.1755, "step": 828 }, { "epoch": 14.310344827586206, "grad_norm": 12.874805450439453, "learning_rate": 5.810983397190294e-05, "loss": 0.4218, "step": 830 }, { "epoch": 14.344827586206897, "grad_norm": 0.3466058671474457, "learning_rate": 5.798212005108558e-05, "loss": 0.1017, "step": 832 }, { "epoch": 14.379310344827585, "grad_norm": 0.1405513435602188, "learning_rate": 5.78544061302682e-05, "loss": 0.0431, "step": 834 }, { "epoch": 14.413793103448276, "grad_norm": 2.837080717086792, "learning_rate": 5.7726692209450826e-05, "loss": 0.1162, "step": 836 }, { "epoch": 14.448275862068966, "grad_norm": 0.49279162287712097, "learning_rate": 5.7598978288633465e-05, "loss": 0.094, "step": 838 }, { "epoch": 14.482758620689655, "grad_norm": 3.950249671936035, "learning_rate": 5.747126436781609e-05, "loss": 0.2088, "step": 840 }, { "epoch": 14.517241379310345, "grad_norm": 9.883584022521973, "learning_rate": 5.734355044699873e-05, "loss": 0.2122, "step": 842 }, { "epoch": 14.551724137931034, "grad_norm": 8.051910400390625, "learning_rate": 5.721583652618135e-05, "loss": 0.3278, "step": 844 }, { "epoch": 14.586206896551724, "grad_norm": 7.80661678314209, "learning_rate": 5.708812260536399e-05, "loss": 0.1348, "step": 846 }, { "epoch": 14.620689655172415, "grad_norm": 6.488097190856934, "learning_rate": 5.6960408684546617e-05, "loss": 0.3984, "step": 848 }, { "epoch": 14.655172413793103, "grad_norm": 12.74706745147705, "learning_rate": 5.683269476372924e-05, "loss": 0.3044, "step": 850 }, { "epoch": 14.689655172413794, "grad_norm": 7.486535549163818, "learning_rate": 5.670498084291188e-05, "loss": 0.3548, "step": 852 }, { "epoch": 14.724137931034482, "grad_norm": 4.710317611694336, "learning_rate": 5.6577266922094505e-05, "loss": 0.131, "step": 854 }, { "epoch": 14.758620689655173, "grad_norm": 0.45310112833976746, "learning_rate": 5.644955300127714e-05, "loss": 0.2907, "step": 856 }, { "epoch": 14.793103448275861, "grad_norm": 11.309956550598145, "learning_rate": 5.632183908045977e-05, "loss": 0.3132, "step": 858 }, { "epoch": 14.827586206896552, "grad_norm": 16.222618103027344, "learning_rate": 5.6194125159642407e-05, "loss": 0.3499, "step": 860 }, { "epoch": 14.862068965517242, "grad_norm": 19.321205139160156, "learning_rate": 5.606641123882503e-05, "loss": 1.193, "step": 862 }, { "epoch": 14.89655172413793, "grad_norm": 14.328798294067383, "learning_rate": 5.593869731800766e-05, "loss": 0.1844, "step": 864 }, { "epoch": 14.931034482758621, "grad_norm": 14.36063289642334, "learning_rate": 5.5810983397190295e-05, "loss": 1.2338, "step": 866 }, { "epoch": 14.96551724137931, "grad_norm": 2.4482498168945312, "learning_rate": 5.5683269476372927e-05, "loss": 0.0565, "step": 868 }, { "epoch": 15.0, "grad_norm": 9.643209457397461, "learning_rate": 5.555555555555556e-05, "loss": 0.1869, "step": 870 }, { "epoch": 15.0, "eval_accuracy": 0.6060606060606061, "eval_f1_macro": 0.5145970477437697, "eval_f1_micro": 0.6060606060606061, "eval_f1_weighted": 0.5982909919119969, "eval_loss": 1.2305328845977783, "eval_precision_macro": 0.5032325523840949, "eval_precision_micro": 0.6060606060606061, "eval_precision_weighted": 0.6012548090624109, "eval_recall_macro": 0.5369360902255639, "eval_recall_micro": 0.6060606060606061, "eval_recall_weighted": 0.6060606060606061, "eval_runtime": 1.9435, "eval_samples_per_second": 67.917, "eval_steps_per_second": 8.747, "step": 870 }, { "epoch": 15.03448275862069, "grad_norm": 6.357883453369141, "learning_rate": 5.542784163473819e-05, "loss": 0.3321, "step": 872 }, { "epoch": 15.068965517241379, "grad_norm": 13.48095989227295, "learning_rate": 5.530012771392082e-05, "loss": 0.1638, "step": 874 }, { "epoch": 15.10344827586207, "grad_norm": 9.955169677734375, "learning_rate": 5.517241379310345e-05, "loss": 0.2777, "step": 876 }, { "epoch": 15.137931034482758, "grad_norm": 6.751859664916992, "learning_rate": 5.504469987228608e-05, "loss": 0.1799, "step": 878 }, { "epoch": 15.172413793103448, "grad_norm": 4.653921127319336, "learning_rate": 5.491698595146872e-05, "loss": 0.2553, "step": 880 }, { "epoch": 15.206896551724139, "grad_norm": 5.335418224334717, "learning_rate": 5.478927203065134e-05, "loss": 0.1899, "step": 882 }, { "epoch": 15.241379310344827, "grad_norm": 5.713738441467285, "learning_rate": 5.466155810983398e-05, "loss": 0.4833, "step": 884 }, { "epoch": 15.275862068965518, "grad_norm": 5.765931129455566, "learning_rate": 5.4533844189016605e-05, "loss": 0.1149, "step": 886 }, { "epoch": 15.310344827586206, "grad_norm": 8.66435718536377, "learning_rate": 5.440613026819924e-05, "loss": 0.3092, "step": 888 }, { "epoch": 15.344827586206897, "grad_norm": 2.088322877883911, "learning_rate": 5.427841634738187e-05, "loss": 0.1967, "step": 890 }, { "epoch": 15.379310344827585, "grad_norm": 12.177457809448242, "learning_rate": 5.415070242656451e-05, "loss": 0.1026, "step": 892 }, { "epoch": 15.413793103448276, "grad_norm": 10.64006233215332, "learning_rate": 5.402298850574713e-05, "loss": 0.2939, "step": 894 }, { "epoch": 15.448275862068966, "grad_norm": 18.743099212646484, "learning_rate": 5.3895274584929756e-05, "loss": 0.2609, "step": 896 }, { "epoch": 15.482758620689655, "grad_norm": 10.08409309387207, "learning_rate": 5.3767560664112395e-05, "loss": 0.1403, "step": 898 }, { "epoch": 15.517241379310345, "grad_norm": 10.607285499572754, "learning_rate": 5.363984674329502e-05, "loss": 0.2563, "step": 900 }, { "epoch": 15.551724137931034, "grad_norm": 2.387784004211426, "learning_rate": 5.351213282247766e-05, "loss": 0.0304, "step": 902 }, { "epoch": 15.586206896551724, "grad_norm": 1.1395808458328247, "learning_rate": 5.338441890166028e-05, "loss": 0.0853, "step": 904 }, { "epoch": 15.620689655172415, "grad_norm": 7.147764682769775, "learning_rate": 5.325670498084292e-05, "loss": 0.3563, "step": 906 }, { "epoch": 15.655172413793103, "grad_norm": 4.958852767944336, "learning_rate": 5.3128991060025547e-05, "loss": 0.1273, "step": 908 }, { "epoch": 15.689655172413794, "grad_norm": 6.212413311004639, "learning_rate": 5.300127713920817e-05, "loss": 0.1748, "step": 910 }, { "epoch": 15.724137931034482, "grad_norm": 1.9379197359085083, "learning_rate": 5.287356321839081e-05, "loss": 0.0297, "step": 912 }, { "epoch": 15.758620689655173, "grad_norm": 1.3377583026885986, "learning_rate": 5.2745849297573435e-05, "loss": 0.0356, "step": 914 }, { "epoch": 15.793103448275861, "grad_norm": 0.8058100938796997, "learning_rate": 5.261813537675607e-05, "loss": 0.1674, "step": 916 }, { "epoch": 15.827586206896552, "grad_norm": 9.972665786743164, "learning_rate": 5.24904214559387e-05, "loss": 0.211, "step": 918 }, { "epoch": 15.862068965517242, "grad_norm": 10.261574745178223, "learning_rate": 5.236270753512134e-05, "loss": 0.2202, "step": 920 }, { "epoch": 15.89655172413793, "grad_norm": 4.782658576965332, "learning_rate": 5.223499361430396e-05, "loss": 0.1333, "step": 922 }, { "epoch": 15.931034482758621, "grad_norm": 7.032027244567871, "learning_rate": 5.2107279693486586e-05, "loss": 0.2523, "step": 924 }, { "epoch": 15.96551724137931, "grad_norm": 0.28720977902412415, "learning_rate": 5.1979565772669225e-05, "loss": 0.0323, "step": 926 }, { "epoch": 16.0, "grad_norm": 9.86831283569336, "learning_rate": 5.185185185185185e-05, "loss": 0.1015, "step": 928 }, { "epoch": 16.0, "eval_accuracy": 0.5909090909090909, "eval_f1_macro": 0.501919425368257, "eval_f1_micro": 0.5909090909090909, "eval_f1_weighted": 0.5932466389209485, "eval_loss": 1.3576174974441528, "eval_precision_macro": 0.5439596385486938, "eval_precision_micro": 0.5909090909090909, "eval_precision_weighted": 0.6311543139507142, "eval_recall_macro": 0.4959429824561404, "eval_recall_micro": 0.5909090909090909, "eval_recall_weighted": 0.5909090909090909, "eval_runtime": 1.9838, "eval_samples_per_second": 66.538, "eval_steps_per_second": 8.569, "step": 928 }, { "epoch": 16.03448275862069, "grad_norm": 2.869445323944092, "learning_rate": 5.172413793103449e-05, "loss": 0.3521, "step": 930 }, { "epoch": 16.06896551724138, "grad_norm": 12.4180908203125, "learning_rate": 5.159642401021711e-05, "loss": 0.3301, "step": 932 }, { "epoch": 16.103448275862068, "grad_norm": 3.6048808097839355, "learning_rate": 5.146871008939975e-05, "loss": 0.1339, "step": 934 }, { "epoch": 16.137931034482758, "grad_norm": 0.6583413481712341, "learning_rate": 5.1340996168582377e-05, "loss": 0.3911, "step": 936 }, { "epoch": 16.17241379310345, "grad_norm": 1.6497776508331299, "learning_rate": 5.1213282247765e-05, "loss": 0.2196, "step": 938 }, { "epoch": 16.20689655172414, "grad_norm": 0.1662745475769043, "learning_rate": 5.108556832694764e-05, "loss": 0.0094, "step": 940 }, { "epoch": 16.24137931034483, "grad_norm": 2.7962138652801514, "learning_rate": 5.0957854406130265e-05, "loss": 0.0605, "step": 942 }, { "epoch": 16.275862068965516, "grad_norm": 5.833149433135986, "learning_rate": 5.08301404853129e-05, "loss": 0.2013, "step": 944 }, { "epoch": 16.310344827586206, "grad_norm": 13.730070114135742, "learning_rate": 5.070242656449553e-05, "loss": 0.2639, "step": 946 }, { "epoch": 16.344827586206897, "grad_norm": 5.876251220703125, "learning_rate": 5.057471264367817e-05, "loss": 0.3352, "step": 948 }, { "epoch": 16.379310344827587, "grad_norm": 3.6209003925323486, "learning_rate": 5.044699872286079e-05, "loss": 0.1183, "step": 950 }, { "epoch": 16.413793103448278, "grad_norm": 0.6604540944099426, "learning_rate": 5.031928480204342e-05, "loss": 0.0778, "step": 952 }, { "epoch": 16.448275862068964, "grad_norm": 1.6417039632797241, "learning_rate": 5.0191570881226055e-05, "loss": 0.0523, "step": 954 }, { "epoch": 16.482758620689655, "grad_norm": 4.871641159057617, "learning_rate": 5.0063856960408687e-05, "loss": 0.1488, "step": 956 }, { "epoch": 16.517241379310345, "grad_norm": 0.6309295892715454, "learning_rate": 4.993614303959132e-05, "loss": 0.039, "step": 958 }, { "epoch": 16.551724137931036, "grad_norm": 9.119877815246582, "learning_rate": 4.980842911877395e-05, "loss": 0.1025, "step": 960 }, { "epoch": 16.586206896551722, "grad_norm": 1.3798292875289917, "learning_rate": 4.968071519795658e-05, "loss": 0.0307, "step": 962 }, { "epoch": 16.620689655172413, "grad_norm": 0.3394834101200104, "learning_rate": 4.955300127713921e-05, "loss": 0.04, "step": 964 }, { "epoch": 16.655172413793103, "grad_norm": 0.6206883788108826, "learning_rate": 4.9425287356321845e-05, "loss": 0.1483, "step": 966 }, { "epoch": 16.689655172413794, "grad_norm": 9.715845108032227, "learning_rate": 4.929757343550448e-05, "loss": 0.3286, "step": 968 }, { "epoch": 16.724137931034484, "grad_norm": 1.0101009607315063, "learning_rate": 4.916985951468711e-05, "loss": 0.158, "step": 970 }, { "epoch": 16.75862068965517, "grad_norm": 17.89195442199707, "learning_rate": 4.904214559386973e-05, "loss": 0.0978, "step": 972 }, { "epoch": 16.79310344827586, "grad_norm": 3.084064245223999, "learning_rate": 4.8914431673052365e-05, "loss": 0.0461, "step": 974 }, { "epoch": 16.82758620689655, "grad_norm": 0.12563039362430573, "learning_rate": 4.8786717752234997e-05, "loss": 0.0234, "step": 976 }, { "epoch": 16.862068965517242, "grad_norm": 1.2776066064834595, "learning_rate": 4.865900383141763e-05, "loss": 0.0931, "step": 978 }, { "epoch": 16.896551724137932, "grad_norm": 2.729607343673706, "learning_rate": 4.853128991060026e-05, "loss": 0.0507, "step": 980 }, { "epoch": 16.93103448275862, "grad_norm": 0.35684946179389954, "learning_rate": 4.840357598978289e-05, "loss": 0.1429, "step": 982 }, { "epoch": 16.96551724137931, "grad_norm": 4.096828460693359, "learning_rate": 4.827586206896552e-05, "loss": 0.321, "step": 984 }, { "epoch": 17.0, "grad_norm": 0.25383704900741577, "learning_rate": 4.814814814814815e-05, "loss": 0.3809, "step": 986 }, { "epoch": 17.0, "eval_accuracy": 0.6590909090909091, "eval_f1_macro": 0.5666514454561888, "eval_f1_micro": 0.6590909090909091, "eval_f1_weighted": 0.652650863169293, "eval_loss": 1.2998155355453491, "eval_precision_macro": 0.5827908086311447, "eval_precision_micro": 0.6590909090909091, "eval_precision_weighted": 0.6885351168439403, "eval_recall_macro": 0.5838189223057644, "eval_recall_micro": 0.6590909090909091, "eval_recall_weighted": 0.6590909090909091, "eval_runtime": 2.7743, "eval_samples_per_second": 47.58, "eval_steps_per_second": 6.128, "step": 986 }, { "epoch": 17.03448275862069, "grad_norm": 2.5796058177948, "learning_rate": 4.802043422733078e-05, "loss": 0.0584, "step": 988 }, { "epoch": 17.06896551724138, "grad_norm": 0.15105970203876495, "learning_rate": 4.789272030651341e-05, "loss": 0.0102, "step": 990 }, { "epoch": 17.103448275862068, "grad_norm": 1.5413386821746826, "learning_rate": 4.776500638569604e-05, "loss": 0.039, "step": 992 }, { "epoch": 17.137931034482758, "grad_norm": 20.432376861572266, "learning_rate": 4.7637292464878675e-05, "loss": 0.1994, "step": 994 }, { "epoch": 17.17241379310345, "grad_norm": 13.159724235534668, "learning_rate": 4.7509578544061307e-05, "loss": 0.1881, "step": 996 }, { "epoch": 17.20689655172414, "grad_norm": 0.3454015552997589, "learning_rate": 4.738186462324394e-05, "loss": 0.0048, "step": 998 }, { "epoch": 17.24137931034483, "grad_norm": 20.270763397216797, "learning_rate": 4.725415070242657e-05, "loss": 0.123, "step": 1000 }, { "epoch": 17.275862068965516, "grad_norm": 22.238080978393555, "learning_rate": 4.7126436781609195e-05, "loss": 0.2302, "step": 1002 }, { "epoch": 17.310344827586206, "grad_norm": 2.1673974990844727, "learning_rate": 4.6998722860791827e-05, "loss": 0.0187, "step": 1004 }, { "epoch": 17.344827586206897, "grad_norm": 0.3968747556209564, "learning_rate": 4.687100893997446e-05, "loss": 0.0076, "step": 1006 }, { "epoch": 17.379310344827587, "grad_norm": 0.5099461078643799, "learning_rate": 4.674329501915709e-05, "loss": 0.0134, "step": 1008 }, { "epoch": 17.413793103448278, "grad_norm": 8.667265892028809, "learning_rate": 4.661558109833972e-05, "loss": 0.2712, "step": 1010 }, { "epoch": 17.448275862068964, "grad_norm": 6.667606353759766, "learning_rate": 4.648786717752235e-05, "loss": 0.1044, "step": 1012 }, { "epoch": 17.482758620689655, "grad_norm": 13.070887565612793, "learning_rate": 4.6360153256704985e-05, "loss": 0.3492, "step": 1014 }, { "epoch": 17.517241379310345, "grad_norm": 4.142744064331055, "learning_rate": 4.623243933588761e-05, "loss": 0.0549, "step": 1016 }, { "epoch": 17.551724137931036, "grad_norm": 0.10112213343381882, "learning_rate": 4.610472541507024e-05, "loss": 0.0567, "step": 1018 }, { "epoch": 17.586206896551722, "grad_norm": 0.03555409982800484, "learning_rate": 4.597701149425287e-05, "loss": 0.0382, "step": 1020 }, { "epoch": 17.620689655172413, "grad_norm": 0.5085342526435852, "learning_rate": 4.5849297573435505e-05, "loss": 0.0261, "step": 1022 }, { "epoch": 17.655172413793103, "grad_norm": 1.362773060798645, "learning_rate": 4.5721583652618137e-05, "loss": 0.1, "step": 1024 }, { "epoch": 17.689655172413794, "grad_norm": 2.903794050216675, "learning_rate": 4.559386973180077e-05, "loss": 0.0157, "step": 1026 }, { "epoch": 17.724137931034484, "grad_norm": 0.29353153705596924, "learning_rate": 4.54661558109834e-05, "loss": 0.0185, "step": 1028 }, { "epoch": 17.75862068965517, "grad_norm": 4.77042293548584, "learning_rate": 4.5338441890166025e-05, "loss": 0.0405, "step": 1030 }, { "epoch": 17.79310344827586, "grad_norm": 2.864474296569824, "learning_rate": 4.5210727969348656e-05, "loss": 0.3131, "step": 1032 }, { "epoch": 17.82758620689655, "grad_norm": 1.4102516174316406, "learning_rate": 4.508301404853129e-05, "loss": 0.1175, "step": 1034 }, { "epoch": 17.862068965517242, "grad_norm": 5.741951942443848, "learning_rate": 4.495530012771392e-05, "loss": 0.3193, "step": 1036 }, { "epoch": 17.896551724137932, "grad_norm": 0.3935476243495941, "learning_rate": 4.482758620689655e-05, "loss": 0.0074, "step": 1038 }, { "epoch": 17.93103448275862, "grad_norm": 6.773815155029297, "learning_rate": 4.469987228607918e-05, "loss": 0.0759, "step": 1040 }, { "epoch": 17.96551724137931, "grad_norm": 14.86123275756836, "learning_rate": 4.4572158365261815e-05, "loss": 0.1914, "step": 1042 }, { "epoch": 18.0, "grad_norm": 5.217438697814941, "learning_rate": 4.4444444444444447e-05, "loss": 0.0887, "step": 1044 }, { "epoch": 18.0, "eval_accuracy": 0.6666666666666666, "eval_f1_macro": 0.5571781194602418, "eval_f1_micro": 0.6666666666666666, "eval_f1_weighted": 0.6489059749380235, "eval_loss": 1.4153717756271362, "eval_precision_macro": 0.5682278640899331, "eval_precision_micro": 0.6666666666666666, "eval_precision_weighted": 0.6518116175012726, "eval_recall_macro": 0.568342731829574, "eval_recall_micro": 0.6666666666666666, "eval_recall_weighted": 0.6666666666666666, "eval_runtime": 2.0979, "eval_samples_per_second": 62.919, "eval_steps_per_second": 8.103, "step": 1044 }, { "epoch": 18.03448275862069, "grad_norm": 15.641746520996094, "learning_rate": 4.431673052362708e-05, "loss": 0.1409, "step": 1046 }, { "epoch": 18.06896551724138, "grad_norm": 11.60824203491211, "learning_rate": 4.418901660280971e-05, "loss": 0.4299, "step": 1048 }, { "epoch": 18.103448275862068, "grad_norm": 0.08567314594984055, "learning_rate": 4.406130268199234e-05, "loss": 0.0062, "step": 1050 }, { "epoch": 18.137931034482758, "grad_norm": 0.12726525962352753, "learning_rate": 4.393358876117497e-05, "loss": 0.0031, "step": 1052 }, { "epoch": 18.17241379310345, "grad_norm": 0.3815951645374298, "learning_rate": 4.3805874840357605e-05, "loss": 0.004, "step": 1054 }, { "epoch": 18.20689655172414, "grad_norm": 17.277345657348633, "learning_rate": 4.367816091954024e-05, "loss": 0.1928, "step": 1056 }, { "epoch": 18.24137931034483, "grad_norm": 0.14995414018630981, "learning_rate": 4.355044699872286e-05, "loss": 0.1884, "step": 1058 }, { "epoch": 18.275862068965516, "grad_norm": 0.1205284371972084, "learning_rate": 4.342273307790549e-05, "loss": 0.1822, "step": 1060 }, { "epoch": 18.310344827586206, "grad_norm": 0.35154199600219727, "learning_rate": 4.3295019157088125e-05, "loss": 0.019, "step": 1062 }, { "epoch": 18.344827586206897, "grad_norm": 4.302131652832031, "learning_rate": 4.3167305236270757e-05, "loss": 0.0265, "step": 1064 }, { "epoch": 18.379310344827587, "grad_norm": 1.0111949443817139, "learning_rate": 4.303959131545339e-05, "loss": 0.0126, "step": 1066 }, { "epoch": 18.413793103448278, "grad_norm": 12.860360145568848, "learning_rate": 4.291187739463602e-05, "loss": 0.1918, "step": 1068 }, { "epoch": 18.448275862068964, "grad_norm": 1.465510368347168, "learning_rate": 4.278416347381865e-05, "loss": 0.1672, "step": 1070 }, { "epoch": 18.482758620689655, "grad_norm": 0.1530478149652481, "learning_rate": 4.2656449553001277e-05, "loss": 0.0624, "step": 1072 }, { "epoch": 18.517241379310345, "grad_norm": 7.125781059265137, "learning_rate": 4.252873563218391e-05, "loss": 0.359, "step": 1074 }, { "epoch": 18.551724137931036, "grad_norm": 3.524441957473755, "learning_rate": 4.240102171136654e-05, "loss": 0.0263, "step": 1076 }, { "epoch": 18.586206896551722, "grad_norm": 2.0102035999298096, "learning_rate": 4.227330779054917e-05, "loss": 0.0602, "step": 1078 }, { "epoch": 18.620689655172413, "grad_norm": 0.3677597939968109, "learning_rate": 4.21455938697318e-05, "loss": 0.0251, "step": 1080 }, { "epoch": 18.655172413793103, "grad_norm": 19.12733268737793, "learning_rate": 4.2017879948914435e-05, "loss": 0.4779, "step": 1082 }, { "epoch": 18.689655172413794, "grad_norm": 2.7869174480438232, "learning_rate": 4.189016602809707e-05, "loss": 0.1591, "step": 1084 }, { "epoch": 18.724137931034484, "grad_norm": 2.577024459838867, "learning_rate": 4.17624521072797e-05, "loss": 0.0862, "step": 1086 }, { "epoch": 18.75862068965517, "grad_norm": 0.28967535495758057, "learning_rate": 4.163473818646232e-05, "loss": 0.0645, "step": 1088 }, { "epoch": 18.79310344827586, "grad_norm": 15.79033088684082, "learning_rate": 4.1507024265644955e-05, "loss": 0.1082, "step": 1090 }, { "epoch": 18.82758620689655, "grad_norm": 0.5589704513549805, "learning_rate": 4.1379310344827587e-05, "loss": 0.0205, "step": 1092 }, { "epoch": 18.862068965517242, "grad_norm": 0.6776816844940186, "learning_rate": 4.125159642401022e-05, "loss": 0.0067, "step": 1094 }, { "epoch": 18.896551724137932, "grad_norm": 0.27798280119895935, "learning_rate": 4.112388250319285e-05, "loss": 0.0039, "step": 1096 }, { "epoch": 18.93103448275862, "grad_norm": 12.297942161560059, "learning_rate": 4.099616858237548e-05, "loss": 0.151, "step": 1098 }, { "epoch": 18.96551724137931, "grad_norm": 2.289623737335205, "learning_rate": 4.086845466155811e-05, "loss": 0.176, "step": 1100 }, { "epoch": 19.0, "grad_norm": 11.227123260498047, "learning_rate": 4.074074074074074e-05, "loss": 0.1422, "step": 1102 }, { "epoch": 19.0, "eval_accuracy": 0.6666666666666666, "eval_f1_macro": 0.5608736126029359, "eval_f1_micro": 0.6666666666666666, "eval_f1_weighted": 0.647220323536113, "eval_loss": 1.398878574371338, "eval_precision_macro": 0.5671867606078133, "eval_precision_micro": 0.6666666666666666, "eval_precision_weighted": 0.6419607605313348, "eval_recall_macro": 0.5695332080200501, "eval_recall_micro": 0.6666666666666666, "eval_recall_weighted": 0.6666666666666666, "eval_runtime": 3.3441, "eval_samples_per_second": 39.472, "eval_steps_per_second": 5.084, "step": 1102 }, { "epoch": 19.03448275862069, "grad_norm": 3.0859837532043457, "learning_rate": 4.061302681992337e-05, "loss": 0.0213, "step": 1104 }, { "epoch": 19.06896551724138, "grad_norm": 12.9458589553833, "learning_rate": 4.0485312899106e-05, "loss": 0.1051, "step": 1106 }, { "epoch": 19.103448275862068, "grad_norm": 0.20263761281967163, "learning_rate": 4.035759897828863e-05, "loss": 0.0172, "step": 1108 }, { "epoch": 19.137931034482758, "grad_norm": 0.19154572486877441, "learning_rate": 4.0229885057471265e-05, "loss": 0.035, "step": 1110 }, { "epoch": 19.17241379310345, "grad_norm": 0.30079570412635803, "learning_rate": 4.0102171136653897e-05, "loss": 0.0363, "step": 1112 }, { "epoch": 19.20689655172414, "grad_norm": 0.17321471869945526, "learning_rate": 3.997445721583653e-05, "loss": 0.0143, "step": 1114 }, { "epoch": 19.24137931034483, "grad_norm": 0.6283549666404724, "learning_rate": 3.984674329501916e-05, "loss": 0.0288, "step": 1116 }, { "epoch": 19.275862068965516, "grad_norm": 0.014564587734639645, "learning_rate": 3.971902937420179e-05, "loss": 0.005, "step": 1118 }, { "epoch": 19.310344827586206, "grad_norm": 0.0865383893251419, "learning_rate": 3.959131545338442e-05, "loss": 0.0134, "step": 1120 }, { "epoch": 19.344827586206897, "grad_norm": 0.13445672392845154, "learning_rate": 3.9463601532567055e-05, "loss": 0.0104, "step": 1122 }, { "epoch": 19.379310344827587, "grad_norm": 1.2037104368209839, "learning_rate": 3.933588761174969e-05, "loss": 0.0126, "step": 1124 }, { "epoch": 19.413793103448278, "grad_norm": 0.8391666412353516, "learning_rate": 3.920817369093232e-05, "loss": 0.0176, "step": 1126 }, { "epoch": 19.448275862068964, "grad_norm": 11.825739860534668, "learning_rate": 3.908045977011495e-05, "loss": 0.0852, "step": 1128 }, { "epoch": 19.482758620689655, "grad_norm": 4.42478609085083, "learning_rate": 3.8952745849297575e-05, "loss": 0.077, "step": 1130 }, { "epoch": 19.517241379310345, "grad_norm": 0.5317286849021912, "learning_rate": 3.8825031928480207e-05, "loss": 0.0204, "step": 1132 }, { "epoch": 19.551724137931036, "grad_norm": 0.39885902404785156, "learning_rate": 3.869731800766284e-05, "loss": 0.0064, "step": 1134 }, { "epoch": 19.586206896551722, "grad_norm": 0.8824494481086731, "learning_rate": 3.856960408684547e-05, "loss": 0.0184, "step": 1136 }, { "epoch": 19.620689655172413, "grad_norm": 1.4008420705795288, "learning_rate": 3.84418901660281e-05, "loss": 0.0778, "step": 1138 }, { "epoch": 19.655172413793103, "grad_norm": 0.5302988290786743, "learning_rate": 3.831417624521073e-05, "loss": 0.0935, "step": 1140 }, { "epoch": 19.689655172413794, "grad_norm": 12.098783493041992, "learning_rate": 3.8186462324393365e-05, "loss": 0.4228, "step": 1142 }, { "epoch": 19.724137931034484, "grad_norm": 2.2826879024505615, "learning_rate": 3.805874840357599e-05, "loss": 0.0317, "step": 1144 }, { "epoch": 19.75862068965517, "grad_norm": 0.4077673852443695, "learning_rate": 3.793103448275862e-05, "loss": 0.0078, "step": 1146 }, { "epoch": 19.79310344827586, "grad_norm": 6.110368728637695, "learning_rate": 3.780332056194125e-05, "loss": 0.1554, "step": 1148 }, { "epoch": 19.82758620689655, "grad_norm": 0.027513781562447548, "learning_rate": 3.7675606641123885e-05, "loss": 0.0625, "step": 1150 }, { "epoch": 19.862068965517242, "grad_norm": 6.546108245849609, "learning_rate": 3.7547892720306517e-05, "loss": 0.0988, "step": 1152 }, { "epoch": 19.896551724137932, "grad_norm": 0.08968118578195572, "learning_rate": 3.742017879948915e-05, "loss": 0.002, "step": 1154 }, { "epoch": 19.93103448275862, "grad_norm": 0.14957240223884583, "learning_rate": 3.729246487867178e-05, "loss": 0.0024, "step": 1156 }, { "epoch": 19.96551724137931, "grad_norm": 21.937902450561523, "learning_rate": 3.716475095785441e-05, "loss": 0.4017, "step": 1158 }, { "epoch": 20.0, "grad_norm": 0.162915900349617, "learning_rate": 3.7037037037037037e-05, "loss": 0.0037, "step": 1160 }, { "epoch": 20.0, "eval_accuracy": 0.6212121212121212, "eval_f1_macro": 0.5241622523135129, "eval_f1_micro": 0.6212121212121212, "eval_f1_weighted": 0.6077529051459533, "eval_loss": 1.5133600234985352, "eval_precision_macro": 0.5262823725981621, "eval_precision_micro": 0.6212121212121212, "eval_precision_weighted": 0.6092636895268474, "eval_recall_macro": 0.537390350877193, "eval_recall_micro": 0.6212121212121212, "eval_recall_weighted": 0.6212121212121212, "eval_runtime": 2.2507, "eval_samples_per_second": 58.65, "eval_steps_per_second": 7.553, "step": 1160 }, { "epoch": 20.03448275862069, "grad_norm": 0.17806491255760193, "learning_rate": 3.690932311621967e-05, "loss": 0.0142, "step": 1162 }, { "epoch": 20.06896551724138, "grad_norm": 0.30130231380462646, "learning_rate": 3.67816091954023e-05, "loss": 0.0266, "step": 1164 }, { "epoch": 20.103448275862068, "grad_norm": 0.30467385053634644, "learning_rate": 3.665389527458493e-05, "loss": 0.0289, "step": 1166 }, { "epoch": 20.137931034482758, "grad_norm": 6.654253959655762, "learning_rate": 3.652618135376756e-05, "loss": 0.1494, "step": 1168 }, { "epoch": 20.17241379310345, "grad_norm": 29.355484008789062, "learning_rate": 3.6398467432950195e-05, "loss": 0.3555, "step": 1170 }, { "epoch": 20.20689655172414, "grad_norm": 0.25190678238868713, "learning_rate": 3.627075351213283e-05, "loss": 0.0065, "step": 1172 }, { "epoch": 20.24137931034483, "grad_norm": 7.903636455535889, "learning_rate": 3.614303959131545e-05, "loss": 0.0348, "step": 1174 }, { "epoch": 20.275862068965516, "grad_norm": 1.2426676750183105, "learning_rate": 3.601532567049808e-05, "loss": 0.0348, "step": 1176 }, { "epoch": 20.310344827586206, "grad_norm": 0.05831276252865791, "learning_rate": 3.5887611749680715e-05, "loss": 0.0171, "step": 1178 }, { "epoch": 20.344827586206897, "grad_norm": 3.8727033138275146, "learning_rate": 3.5759897828863347e-05, "loss": 0.2391, "step": 1180 }, { "epoch": 20.379310344827587, "grad_norm": 2.7642955780029297, "learning_rate": 3.563218390804598e-05, "loss": 0.0218, "step": 1182 }, { "epoch": 20.413793103448278, "grad_norm": 28.808361053466797, "learning_rate": 3.550446998722861e-05, "loss": 0.1445, "step": 1184 }, { "epoch": 20.448275862068964, "grad_norm": 0.25608864426612854, "learning_rate": 3.537675606641124e-05, "loss": 0.0038, "step": 1186 }, { "epoch": 20.482758620689655, "grad_norm": 0.18029409646987915, "learning_rate": 3.5249042145593867e-05, "loss": 0.0034, "step": 1188 }, { "epoch": 20.517241379310345, "grad_norm": 0.03822282329201698, "learning_rate": 3.51213282247765e-05, "loss": 0.0623, "step": 1190 }, { "epoch": 20.551724137931036, "grad_norm": 0.24421456456184387, "learning_rate": 3.499361430395913e-05, "loss": 0.0054, "step": 1192 }, { "epoch": 20.586206896551722, "grad_norm": 13.820355415344238, "learning_rate": 3.486590038314176e-05, "loss": 0.0834, "step": 1194 }, { "epoch": 20.620689655172413, "grad_norm": 3.458310127258301, "learning_rate": 3.473818646232439e-05, "loss": 0.0236, "step": 1196 }, { "epoch": 20.655172413793103, "grad_norm": 0.02704853191971779, "learning_rate": 3.4610472541507025e-05, "loss": 0.0033, "step": 1198 }, { "epoch": 20.689655172413794, "grad_norm": 1.1011626720428467, "learning_rate": 3.4482758620689657e-05, "loss": 0.0165, "step": 1200 }, { "epoch": 20.724137931034484, "grad_norm": 21.770254135131836, "learning_rate": 3.435504469987229e-05, "loss": 0.1164, "step": 1202 }, { "epoch": 20.75862068965517, "grad_norm": 0.3167782425880432, "learning_rate": 3.422733077905492e-05, "loss": 0.0036, "step": 1204 }, { "epoch": 20.79310344827586, "grad_norm": 0.057486578822135925, "learning_rate": 3.409961685823755e-05, "loss": 0.0029, "step": 1206 }, { "epoch": 20.82758620689655, "grad_norm": 0.7752865552902222, "learning_rate": 3.397190293742018e-05, "loss": 0.1632, "step": 1208 }, { "epoch": 20.862068965517242, "grad_norm": 1.2659624814987183, "learning_rate": 3.3844189016602815e-05, "loss": 0.0084, "step": 1210 }, { "epoch": 20.896551724137932, "grad_norm": 0.0808514654636383, "learning_rate": 3.371647509578545e-05, "loss": 0.0011, "step": 1212 }, { "epoch": 20.93103448275862, "grad_norm": 0.502754807472229, "learning_rate": 3.358876117496808e-05, "loss": 0.0062, "step": 1214 }, { "epoch": 20.96551724137931, "grad_norm": 0.09211990982294083, "learning_rate": 3.34610472541507e-05, "loss": 0.0043, "step": 1216 }, { "epoch": 21.0, "grad_norm": 0.22179439663887024, "learning_rate": 3.3333333333333335e-05, "loss": 0.0602, "step": 1218 }, { "epoch": 21.0, "eval_accuracy": 0.6666666666666666, "eval_f1_macro": 0.5660444695559945, "eval_f1_micro": 0.6666666666666666, "eval_f1_weighted": 0.6544395105724697, "eval_loss": 1.5348953008651733, "eval_precision_macro": 0.5710371698009604, "eval_precision_micro": 0.6666666666666666, "eval_precision_weighted": 0.6503411219319607, "eval_recall_macro": 0.5670739348370927, "eval_recall_micro": 0.6666666666666666, "eval_recall_weighted": 0.6666666666666666, "eval_runtime": 2.9914, "eval_samples_per_second": 44.127, "eval_steps_per_second": 5.683, "step": 1218 }, { "epoch": 21.03448275862069, "grad_norm": 0.02936175838112831, "learning_rate": 3.3205619412515967e-05, "loss": 0.0007, "step": 1220 }, { "epoch": 21.06896551724138, "grad_norm": 10.468208312988281, "learning_rate": 3.30779054916986e-05, "loss": 0.0762, "step": 1222 }, { "epoch": 21.103448275862068, "grad_norm": 0.010613945312798023, "learning_rate": 3.295019157088123e-05, "loss": 0.0008, "step": 1224 }, { "epoch": 21.137931034482758, "grad_norm": 0.8747249841690063, "learning_rate": 3.282247765006386e-05, "loss": 0.3968, "step": 1226 }, { "epoch": 21.17241379310345, "grad_norm": 0.014937883242964745, "learning_rate": 3.269476372924649e-05, "loss": 0.0526, "step": 1228 }, { "epoch": 21.20689655172414, "grad_norm": 0.29066944122314453, "learning_rate": 3.256704980842912e-05, "loss": 0.0051, "step": 1230 }, { "epoch": 21.24137931034483, "grad_norm": 0.05553491786122322, "learning_rate": 3.243933588761175e-05, "loss": 0.0015, "step": 1232 }, { "epoch": 21.275862068965516, "grad_norm": 0.1087840124964714, "learning_rate": 3.231162196679438e-05, "loss": 0.0035, "step": 1234 }, { "epoch": 21.310344827586206, "grad_norm": 2.526913642883301, "learning_rate": 3.218390804597701e-05, "loss": 0.0228, "step": 1236 }, { "epoch": 21.344827586206897, "grad_norm": 0.09952912479639053, "learning_rate": 3.2056194125159645e-05, "loss": 0.0025, "step": 1238 }, { "epoch": 21.379310344827587, "grad_norm": 0.03265494853258133, "learning_rate": 3.192848020434228e-05, "loss": 0.0044, "step": 1240 }, { "epoch": 21.413793103448278, "grad_norm": 5.046082973480225, "learning_rate": 3.180076628352491e-05, "loss": 0.058, "step": 1242 }, { "epoch": 21.448275862068964, "grad_norm": 15.606002807617188, "learning_rate": 3.167305236270754e-05, "loss": 0.1473, "step": 1244 }, { "epoch": 21.482758620689655, "grad_norm": 19.64705467224121, "learning_rate": 3.1545338441890165e-05, "loss": 0.1913, "step": 1246 }, { "epoch": 21.517241379310345, "grad_norm": 4.021636009216309, "learning_rate": 3.1417624521072797e-05, "loss": 0.0729, "step": 1248 }, { "epoch": 21.551724137931036, "grad_norm": 0.22693483531475067, "learning_rate": 3.128991060025543e-05, "loss": 0.0402, "step": 1250 }, { "epoch": 21.586206896551722, "grad_norm": 0.1609225869178772, "learning_rate": 3.116219667943806e-05, "loss": 0.0025, "step": 1252 }, { "epoch": 21.620689655172413, "grad_norm": 0.10900776833295822, "learning_rate": 3.103448275862069e-05, "loss": 0.014, "step": 1254 }, { "epoch": 21.655172413793103, "grad_norm": 0.47104957699775696, "learning_rate": 3.090676883780332e-05, "loss": 0.0057, "step": 1256 }, { "epoch": 21.689655172413794, "grad_norm": 2.721670150756836, "learning_rate": 3.0779054916985955e-05, "loss": 0.0277, "step": 1258 }, { "epoch": 21.724137931034484, "grad_norm": 0.09523604065179825, "learning_rate": 3.065134099616858e-05, "loss": 0.0015, "step": 1260 }, { "epoch": 21.75862068965517, "grad_norm": 4.91738748550415, "learning_rate": 3.052362707535121e-05, "loss": 0.0663, "step": 1262 }, { "epoch": 21.79310344827586, "grad_norm": 0.12091673910617828, "learning_rate": 3.0395913154533843e-05, "loss": 0.0024, "step": 1264 }, { "epoch": 21.82758620689655, "grad_norm": 11.164044380187988, "learning_rate": 3.0268199233716475e-05, "loss": 0.0475, "step": 1266 }, { "epoch": 21.862068965517242, "grad_norm": 0.018948553130030632, "learning_rate": 3.0140485312899107e-05, "loss": 0.0597, "step": 1268 }, { "epoch": 21.896551724137932, "grad_norm": 4.384052753448486, "learning_rate": 3.0012771392081738e-05, "loss": 0.1747, "step": 1270 }, { "epoch": 21.93103448275862, "grad_norm": 8.05239486694336, "learning_rate": 2.988505747126437e-05, "loss": 0.0828, "step": 1272 }, { "epoch": 21.96551724137931, "grad_norm": 0.07449853420257568, "learning_rate": 2.9757343550446998e-05, "loss": 0.0028, "step": 1274 }, { "epoch": 22.0, "grad_norm": 6.600012302398682, "learning_rate": 2.962962962962963e-05, "loss": 0.0353, "step": 1276 }, { "epoch": 22.0, "eval_accuracy": 0.7045454545454546, "eval_f1_macro": 0.6136674105378829, "eval_f1_micro": 0.7045454545454546, "eval_f1_weighted": 0.6918989425351217, "eval_loss": 1.448930263519287, "eval_precision_macro": 0.6146465063758297, "eval_precision_micro": 0.7045454545454546, "eval_precision_weighted": 0.6908581184896975, "eval_recall_macro": 0.6242167919799498, "eval_recall_micro": 0.7045454545454546, "eval_recall_weighted": 0.7045454545454546, "eval_runtime": 2.3557, "eval_samples_per_second": 56.034, "eval_steps_per_second": 7.217, "step": 1276 }, { "epoch": 22.03448275862069, "grad_norm": 0.022473517805337906, "learning_rate": 2.950191570881226e-05, "loss": 0.0027, "step": 1278 }, { "epoch": 22.06896551724138, "grad_norm": 0.06254889816045761, "learning_rate": 2.9374201787994893e-05, "loss": 0.0087, "step": 1280 }, { "epoch": 22.103448275862068, "grad_norm": 0.07748962938785553, "learning_rate": 2.9246487867177525e-05, "loss": 0.0042, "step": 1282 }, { "epoch": 22.137931034482758, "grad_norm": 0.07348134368658066, "learning_rate": 2.9118773946360157e-05, "loss": 0.0018, "step": 1284 }, { "epoch": 22.17241379310345, "grad_norm": 5.811524868011475, "learning_rate": 2.899106002554279e-05, "loss": 0.0482, "step": 1286 }, { "epoch": 22.20689655172414, "grad_norm": 0.06995466351509094, "learning_rate": 2.8863346104725413e-05, "loss": 0.002, "step": 1288 }, { "epoch": 22.24137931034483, "grad_norm": 0.033994317054748535, "learning_rate": 2.8735632183908045e-05, "loss": 0.0219, "step": 1290 }, { "epoch": 22.275862068965516, "grad_norm": 0.03432360664010048, "learning_rate": 2.8607918263090677e-05, "loss": 0.0009, "step": 1292 }, { "epoch": 22.310344827586206, "grad_norm": 0.01812071166932583, "learning_rate": 2.8480204342273308e-05, "loss": 0.0015, "step": 1294 }, { "epoch": 22.344827586206897, "grad_norm": 1.0397570133209229, "learning_rate": 2.835249042145594e-05, "loss": 0.0078, "step": 1296 }, { "epoch": 22.379310344827587, "grad_norm": 0.5670585036277771, "learning_rate": 2.822477650063857e-05, "loss": 0.0359, "step": 1298 }, { "epoch": 22.413793103448278, "grad_norm": 0.03518284112215042, "learning_rate": 2.8097062579821203e-05, "loss": 0.001, "step": 1300 }, { "epoch": 22.448275862068964, "grad_norm": 0.3894251585006714, "learning_rate": 2.796934865900383e-05, "loss": 0.0021, "step": 1302 }, { "epoch": 22.482758620689655, "grad_norm": 0.12059098482131958, "learning_rate": 2.7841634738186463e-05, "loss": 0.0014, "step": 1304 }, { "epoch": 22.517241379310345, "grad_norm": 3.6728742122650146, "learning_rate": 2.7713920817369095e-05, "loss": 0.0199, "step": 1306 }, { "epoch": 22.551724137931036, "grad_norm": 0.029478440061211586, "learning_rate": 2.7586206896551727e-05, "loss": 0.0009, "step": 1308 }, { "epoch": 22.586206896551722, "grad_norm": 0.31572234630584717, "learning_rate": 2.745849297573436e-05, "loss": 0.004, "step": 1310 }, { "epoch": 22.620689655172413, "grad_norm": 0.04008958116173744, "learning_rate": 2.733077905491699e-05, "loss": 0.0017, "step": 1312 }, { "epoch": 22.655172413793103, "grad_norm": 1.0323200225830078, "learning_rate": 2.720306513409962e-05, "loss": 0.0077, "step": 1314 }, { "epoch": 22.689655172413794, "grad_norm": 0.032519642263650894, "learning_rate": 2.7075351213282253e-05, "loss": 0.0027, "step": 1316 }, { "epoch": 22.724137931034484, "grad_norm": 2.1410605907440186, "learning_rate": 2.6947637292464878e-05, "loss": 0.0125, "step": 1318 }, { "epoch": 22.75862068965517, "grad_norm": 0.016050921753048897, "learning_rate": 2.681992337164751e-05, "loss": 0.0331, "step": 1320 }, { "epoch": 22.79310344827586, "grad_norm": 0.1428113877773285, "learning_rate": 2.669220945083014e-05, "loss": 0.0049, "step": 1322 }, { "epoch": 22.82758620689655, "grad_norm": 0.1641322523355484, "learning_rate": 2.6564495530012773e-05, "loss": 0.002, "step": 1324 }, { "epoch": 22.862068965517242, "grad_norm": 7.986566066741943, "learning_rate": 2.6436781609195405e-05, "loss": 0.0351, "step": 1326 }, { "epoch": 22.896551724137932, "grad_norm": 0.08804414421319962, "learning_rate": 2.6309067688378037e-05, "loss": 0.0013, "step": 1328 }, { "epoch": 22.93103448275862, "grad_norm": 7.736631870269775, "learning_rate": 2.618135376756067e-05, "loss": 0.0779, "step": 1330 }, { "epoch": 22.96551724137931, "grad_norm": 0.015013402327895164, "learning_rate": 2.6053639846743293e-05, "loss": 0.0827, "step": 1332 }, { "epoch": 23.0, "grad_norm": 0.01156703382730484, "learning_rate": 2.5925925925925925e-05, "loss": 0.001, "step": 1334 }, { "epoch": 23.0, "eval_accuracy": 0.6666666666666666, "eval_f1_macro": 0.5714664821494502, "eval_f1_micro": 0.6666666666666666, "eval_f1_weighted": 0.6541314709388933, "eval_loss": 1.4781274795532227, "eval_precision_macro": 0.5657452925012241, "eval_precision_micro": 0.6666666666666666, "eval_precision_weighted": 0.6448584800964208, "eval_recall_macro": 0.580545112781955, "eval_recall_micro": 0.6666666666666666, "eval_recall_weighted": 0.6666666666666666, "eval_runtime": 1.954, "eval_samples_per_second": 67.553, "eval_steps_per_second": 8.7, "step": 1334 }, { "epoch": 23.03448275862069, "grad_norm": 0.006409560330212116, "learning_rate": 2.5798212005108557e-05, "loss": 0.0018, "step": 1336 }, { "epoch": 23.06896551724138, "grad_norm": 2.2381739616394043, "learning_rate": 2.5670498084291188e-05, "loss": 0.0467, "step": 1338 }, { "epoch": 23.103448275862068, "grad_norm": 0.5874573588371277, "learning_rate": 2.554278416347382e-05, "loss": 0.0048, "step": 1340 }, { "epoch": 23.137931034482758, "grad_norm": 22.487873077392578, "learning_rate": 2.541507024265645e-05, "loss": 0.1, "step": 1342 }, { "epoch": 23.17241379310345, "grad_norm": 0.020600179210305214, "learning_rate": 2.5287356321839083e-05, "loss": 0.025, "step": 1344 }, { "epoch": 23.20689655172414, "grad_norm": 0.17713558673858643, "learning_rate": 2.515964240102171e-05, "loss": 0.0037, "step": 1346 }, { "epoch": 23.24137931034483, "grad_norm": 0.19636306166648865, "learning_rate": 2.5031928480204343e-05, "loss": 0.0024, "step": 1348 }, { "epoch": 23.275862068965516, "grad_norm": 0.5884350538253784, "learning_rate": 2.4904214559386975e-05, "loss": 0.0041, "step": 1350 }, { "epoch": 23.310344827586206, "grad_norm": 0.07146797329187393, "learning_rate": 2.4776500638569607e-05, "loss": 0.0012, "step": 1352 }, { "epoch": 23.344827586206897, "grad_norm": 0.11290651559829712, "learning_rate": 2.464878671775224e-05, "loss": 0.008, "step": 1354 }, { "epoch": 23.379310344827587, "grad_norm": 0.09537604451179504, "learning_rate": 2.4521072796934867e-05, "loss": 0.0339, "step": 1356 }, { "epoch": 23.413793103448278, "grad_norm": 0.27110394835472107, "learning_rate": 2.4393358876117498e-05, "loss": 0.0021, "step": 1358 }, { "epoch": 23.448275862068964, "grad_norm": 0.2351570874452591, "learning_rate": 2.426564495530013e-05, "loss": 0.0015, "step": 1360 }, { "epoch": 23.482758620689655, "grad_norm": 0.018089979887008667, "learning_rate": 2.413793103448276e-05, "loss": 0.0023, "step": 1362 }, { "epoch": 23.517241379310345, "grad_norm": 0.12375891953706741, "learning_rate": 2.401021711366539e-05, "loss": 0.0035, "step": 1364 }, { "epoch": 23.551724137931036, "grad_norm": 0.2663002908229828, "learning_rate": 2.388250319284802e-05, "loss": 0.0048, "step": 1366 }, { "epoch": 23.586206896551722, "grad_norm": 0.031247558072209358, "learning_rate": 2.3754789272030653e-05, "loss": 0.092, "step": 1368 }, { "epoch": 23.620689655172413, "grad_norm": 0.024893470108509064, "learning_rate": 2.3627075351213285e-05, "loss": 0.0008, "step": 1370 }, { "epoch": 23.655172413793103, "grad_norm": 0.040011290460824966, "learning_rate": 2.3499361430395913e-05, "loss": 0.0017, "step": 1372 }, { "epoch": 23.689655172413794, "grad_norm": 0.6574517488479614, "learning_rate": 2.3371647509578545e-05, "loss": 0.024, "step": 1374 }, { "epoch": 23.724137931034484, "grad_norm": 0.02834693342447281, "learning_rate": 2.3243933588761177e-05, "loss": 0.0006, "step": 1376 }, { "epoch": 23.75862068965517, "grad_norm": 1.0888547897338867, "learning_rate": 2.3116219667943805e-05, "loss": 0.006, "step": 1378 }, { "epoch": 23.79310344827586, "grad_norm": 0.03572368249297142, "learning_rate": 2.2988505747126437e-05, "loss": 0.0028, "step": 1380 }, { "epoch": 23.82758620689655, "grad_norm": 5.217251777648926, "learning_rate": 2.2860791826309068e-05, "loss": 0.1481, "step": 1382 }, { "epoch": 23.862068965517242, "grad_norm": 0.4067634046077728, "learning_rate": 2.27330779054917e-05, "loss": 0.0045, "step": 1384 }, { "epoch": 23.896551724137932, "grad_norm": 0.1273973435163498, "learning_rate": 2.2605363984674328e-05, "loss": 0.0013, "step": 1386 }, { "epoch": 23.93103448275862, "grad_norm": 0.3500019907951355, "learning_rate": 2.247765006385696e-05, "loss": 0.0023, "step": 1388 }, { "epoch": 23.96551724137931, "grad_norm": 2.0916292667388916, "learning_rate": 2.234993614303959e-05, "loss": 0.0074, "step": 1390 }, { "epoch": 24.0, "grad_norm": 0.03820287436246872, "learning_rate": 2.2222222222222223e-05, "loss": 0.0007, "step": 1392 }, { "epoch": 24.0, "eval_accuracy": 0.6590909090909091, "eval_f1_macro": 0.5712607596935955, "eval_f1_micro": 0.6590909090909091, "eval_f1_weighted": 0.6510971028694909, "eval_loss": 1.632594347000122, "eval_precision_macro": 0.5870793772054277, "eval_precision_micro": 0.6590909090909091, "eval_precision_weighted": 0.664782083632351, "eval_recall_macro": 0.5786027568922305, "eval_recall_micro": 0.6590909090909091, "eval_recall_weighted": 0.6590909090909091, "eval_runtime": 1.9749, "eval_samples_per_second": 66.84, "eval_steps_per_second": 8.608, "step": 1392 }, { "epoch": 24.03448275862069, "grad_norm": 0.07503662258386612, "learning_rate": 2.2094508301404855e-05, "loss": 0.0244, "step": 1394 }, { "epoch": 24.06896551724138, "grad_norm": 27.82219123840332, "learning_rate": 2.1966794380587487e-05, "loss": 0.196, "step": 1396 }, { "epoch": 24.103448275862068, "grad_norm": 0.029466254636645317, "learning_rate": 2.183908045977012e-05, "loss": 0.0007, "step": 1398 }, { "epoch": 24.137931034482758, "grad_norm": 0.15405841171741486, "learning_rate": 2.1711366538952747e-05, "loss": 0.0015, "step": 1400 }, { "epoch": 24.17241379310345, "grad_norm": 2.168177366256714, "learning_rate": 2.1583652618135378e-05, "loss": 0.0114, "step": 1402 }, { "epoch": 24.20689655172414, "grad_norm": 1.049078345298767, "learning_rate": 2.145593869731801e-05, "loss": 0.0039, "step": 1404 }, { "epoch": 24.24137931034483, "grad_norm": 0.10657618939876556, "learning_rate": 2.1328224776500638e-05, "loss": 0.0016, "step": 1406 }, { "epoch": 24.275862068965516, "grad_norm": 2.1004295349121094, "learning_rate": 2.120051085568327e-05, "loss": 0.0427, "step": 1408 }, { "epoch": 24.310344827586206, "grad_norm": 0.026559889316558838, "learning_rate": 2.10727969348659e-05, "loss": 0.0008, "step": 1410 }, { "epoch": 24.344827586206897, "grad_norm": 0.2494554966688156, "learning_rate": 2.0945083014048533e-05, "loss": 0.0021, "step": 1412 }, { "epoch": 24.379310344827587, "grad_norm": 0.5512819290161133, "learning_rate": 2.081736909323116e-05, "loss": 0.0093, "step": 1414 }, { "epoch": 24.413793103448278, "grad_norm": 0.032853931188583374, "learning_rate": 2.0689655172413793e-05, "loss": 0.0006, "step": 1416 }, { "epoch": 24.448275862068964, "grad_norm": 0.29787227511405945, "learning_rate": 2.0561941251596425e-05, "loss": 0.0037, "step": 1418 }, { "epoch": 24.482758620689655, "grad_norm": 0.22131195664405823, "learning_rate": 2.0434227330779057e-05, "loss": 0.0086, "step": 1420 }, { "epoch": 24.517241379310345, "grad_norm": 0.15168803930282593, "learning_rate": 2.0306513409961685e-05, "loss": 0.0017, "step": 1422 }, { "epoch": 24.551724137931036, "grad_norm": 0.006662359926849604, "learning_rate": 2.0178799489144317e-05, "loss": 0.1826, "step": 1424 }, { "epoch": 24.586206896551722, "grad_norm": 0.008878910914063454, "learning_rate": 2.0051085568326948e-05, "loss": 0.0096, "step": 1426 }, { "epoch": 24.620689655172413, "grad_norm": 0.09360943734645844, "learning_rate": 1.992337164750958e-05, "loss": 0.0015, "step": 1428 }, { "epoch": 24.655172413793103, "grad_norm": 0.13396118581295013, "learning_rate": 1.979565772669221e-05, "loss": 0.3838, "step": 1430 }, { "epoch": 24.689655172413794, "grad_norm": 0.010010026395320892, "learning_rate": 1.9667943805874843e-05, "loss": 0.0005, "step": 1432 }, { "epoch": 24.724137931034484, "grad_norm": 1.2189490795135498, "learning_rate": 1.9540229885057475e-05, "loss": 0.0052, "step": 1434 }, { "epoch": 24.75862068965517, "grad_norm": 0.0254853293299675, "learning_rate": 1.9412515964240103e-05, "loss": 0.007, "step": 1436 }, { "epoch": 24.79310344827586, "grad_norm": 5.6452765464782715, "learning_rate": 1.9284802043422735e-05, "loss": 0.0186, "step": 1438 }, { "epoch": 24.82758620689655, "grad_norm": 0.12432337552309036, "learning_rate": 1.9157088122605367e-05, "loss": 0.0012, "step": 1440 }, { "epoch": 24.862068965517242, "grad_norm": 0.06803122162818909, "learning_rate": 1.9029374201787995e-05, "loss": 0.001, "step": 1442 }, { "epoch": 24.896551724137932, "grad_norm": 0.07736227661371231, "learning_rate": 1.8901660280970627e-05, "loss": 0.0007, "step": 1444 }, { "epoch": 24.93103448275862, "grad_norm": 0.140510693192482, "learning_rate": 1.8773946360153258e-05, "loss": 0.0032, "step": 1446 }, { "epoch": 24.96551724137931, "grad_norm": 0.1517113447189331, "learning_rate": 1.864623243933589e-05, "loss": 0.0142, "step": 1448 }, { "epoch": 25.0, "grad_norm": 1.449758768081665, "learning_rate": 1.8518518518518518e-05, "loss": 0.0084, "step": 1450 }, { "epoch": 25.0, "eval_accuracy": 0.6590909090909091, "eval_f1_macro": 0.5684048691621123, "eval_f1_micro": 0.6590909090909091, "eval_f1_weighted": 0.6569451587148032, "eval_loss": 1.5855995416641235, "eval_precision_macro": 0.5662025575780552, "eval_precision_micro": 0.6590909090909091, "eval_precision_weighted": 0.6671501716534084, "eval_recall_macro": 0.5802474937343358, "eval_recall_micro": 0.6590909090909091, "eval_recall_weighted": 0.6590909090909091, "eval_runtime": 2.0013, "eval_samples_per_second": 65.957, "eval_steps_per_second": 8.495, "step": 1450 }, { "epoch": 25.03448275862069, "grad_norm": 3.8146300315856934, "learning_rate": 1.839080459770115e-05, "loss": 0.116, "step": 1452 }, { "epoch": 25.06896551724138, "grad_norm": 0.013178830966353416, "learning_rate": 1.826309067688378e-05, "loss": 0.0004, "step": 1454 }, { "epoch": 25.103448275862068, "grad_norm": 0.04289768263697624, "learning_rate": 1.8135376756066413e-05, "loss": 0.001, "step": 1456 }, { "epoch": 25.137931034482758, "grad_norm": 0.3486016094684601, "learning_rate": 1.800766283524904e-05, "loss": 0.0018, "step": 1458 }, { "epoch": 25.17241379310345, "grad_norm": 0.101795494556427, "learning_rate": 1.7879948914431673e-05, "loss": 0.0487, "step": 1460 }, { "epoch": 25.20689655172414, "grad_norm": 0.008328469470143318, "learning_rate": 1.7752234993614305e-05, "loss": 0.0003, "step": 1462 }, { "epoch": 25.24137931034483, "grad_norm": 0.022965148091316223, "learning_rate": 1.7624521072796933e-05, "loss": 0.0007, "step": 1464 }, { "epoch": 25.275862068965516, "grad_norm": 0.05658239498734474, "learning_rate": 1.7496807151979565e-05, "loss": 0.0012, "step": 1466 }, { "epoch": 25.310344827586206, "grad_norm": 0.4014154374599457, "learning_rate": 1.7369093231162197e-05, "loss": 0.0046, "step": 1468 }, { "epoch": 25.344827586206897, "grad_norm": 5.419172763824463, "learning_rate": 1.7241379310344828e-05, "loss": 0.0219, "step": 1470 }, { "epoch": 25.379310344827587, "grad_norm": 0.022299258038401604, "learning_rate": 1.711366538952746e-05, "loss": 0.0007, "step": 1472 }, { "epoch": 25.413793103448278, "grad_norm": 0.8597404360771179, "learning_rate": 1.698595146871009e-05, "loss": 0.0042, "step": 1474 }, { "epoch": 25.448275862068964, "grad_norm": 0.056232478469610214, "learning_rate": 1.6858237547892723e-05, "loss": 0.0007, "step": 1476 }, { "epoch": 25.482758620689655, "grad_norm": 0.8406282663345337, "learning_rate": 1.673052362707535e-05, "loss": 0.0036, "step": 1478 }, { "epoch": 25.517241379310345, "grad_norm": 0.03363404422998428, "learning_rate": 1.6602809706257983e-05, "loss": 0.0014, "step": 1480 }, { "epoch": 25.551724137931036, "grad_norm": 0.015727031975984573, "learning_rate": 1.6475095785440615e-05, "loss": 0.0008, "step": 1482 }, { "epoch": 25.586206896551722, "grad_norm": 0.03139304369688034, "learning_rate": 1.6347381864623247e-05, "loss": 0.0017, "step": 1484 }, { "epoch": 25.620689655172413, "grad_norm": 0.02697753719985485, "learning_rate": 1.6219667943805875e-05, "loss": 0.0008, "step": 1486 }, { "epoch": 25.655172413793103, "grad_norm": 0.004935800563544035, "learning_rate": 1.6091954022988507e-05, "loss": 0.0006, "step": 1488 }, { "epoch": 25.689655172413794, "grad_norm": 1.9771162271499634, "learning_rate": 1.596424010217114e-05, "loss": 0.0139, "step": 1490 }, { "epoch": 25.724137931034484, "grad_norm": 0.014661573804914951, "learning_rate": 1.583652618135377e-05, "loss": 0.0004, "step": 1492 }, { "epoch": 25.75862068965517, "grad_norm": 2.9789557456970215, "learning_rate": 1.5708812260536398e-05, "loss": 0.0784, "step": 1494 }, { "epoch": 25.79310344827586, "grad_norm": 0.05912244692444801, "learning_rate": 1.558109833971903e-05, "loss": 0.002, "step": 1496 }, { "epoch": 25.82758620689655, "grad_norm": 0.21335746347904205, "learning_rate": 1.545338441890166e-05, "loss": 0.002, "step": 1498 }, { "epoch": 25.862068965517242, "grad_norm": 0.01072037685662508, "learning_rate": 1.532567049808429e-05, "loss": 0.0005, "step": 1500 }, { "epoch": 25.896551724137932, "grad_norm": 0.04208362475037575, "learning_rate": 1.5197956577266922e-05, "loss": 0.0007, "step": 1502 }, { "epoch": 25.93103448275862, "grad_norm": 0.026036394760012627, "learning_rate": 1.5070242656449553e-05, "loss": 0.0006, "step": 1504 }, { "epoch": 25.96551724137931, "grad_norm": 0.08341384679079056, "learning_rate": 1.4942528735632185e-05, "loss": 0.0012, "step": 1506 }, { "epoch": 26.0, "grad_norm": 0.0347595140337944, "learning_rate": 1.4814814814814815e-05, "loss": 0.0008, "step": 1508 }, { "epoch": 26.0, "eval_accuracy": 0.6818181818181818, "eval_f1_macro": 0.5826228878648233, "eval_f1_micro": 0.6818181818181818, "eval_f1_weighted": 0.6675219941348973, "eval_loss": 1.5798661708831787, "eval_precision_macro": 0.5849358974358975, "eval_precision_micro": 0.6818181818181818, "eval_precision_weighted": 0.6631993006993008, "eval_recall_macro": 0.5884241854636592, "eval_recall_micro": 0.6818181818181818, "eval_recall_weighted": 0.6818181818181818, "eval_runtime": 1.9704, "eval_samples_per_second": 66.992, "eval_steps_per_second": 8.628, "step": 1508 }, { "epoch": 26.03448275862069, "grad_norm": 0.10061348974704742, "learning_rate": 1.4687100893997447e-05, "loss": 0.0221, "step": 1510 }, { "epoch": 26.06896551724138, "grad_norm": 0.005162129178643227, "learning_rate": 1.4559386973180078e-05, "loss": 0.0003, "step": 1512 }, { "epoch": 26.103448275862068, "grad_norm": 0.02409091778099537, "learning_rate": 1.4431673052362707e-05, "loss": 0.0004, "step": 1514 }, { "epoch": 26.137931034482758, "grad_norm": 0.02003251574933529, "learning_rate": 1.4303959131545338e-05, "loss": 0.0009, "step": 1516 }, { "epoch": 26.17241379310345, "grad_norm": 0.11420497298240662, "learning_rate": 1.417624521072797e-05, "loss": 0.0011, "step": 1518 }, { "epoch": 26.20689655172414, "grad_norm": 0.03824389725923538, "learning_rate": 1.4048531289910602e-05, "loss": 0.0008, "step": 1520 }, { "epoch": 26.24137931034483, "grad_norm": 0.00969379860907793, "learning_rate": 1.3920817369093232e-05, "loss": 0.0003, "step": 1522 }, { "epoch": 26.275862068965516, "grad_norm": 1.241509199142456, "learning_rate": 1.3793103448275863e-05, "loss": 0.0038, "step": 1524 }, { "epoch": 26.310344827586206, "grad_norm": 0.011483917012810707, "learning_rate": 1.3665389527458495e-05, "loss": 0.0011, "step": 1526 }, { "epoch": 26.344827586206897, "grad_norm": 0.02756449580192566, "learning_rate": 1.3537675606641127e-05, "loss": 0.0007, "step": 1528 }, { "epoch": 26.379310344827587, "grad_norm": 0.01528858207166195, "learning_rate": 1.3409961685823755e-05, "loss": 0.0003, "step": 1530 }, { "epoch": 26.413793103448278, "grad_norm": 0.01113022118806839, "learning_rate": 1.3282247765006387e-05, "loss": 0.0005, "step": 1532 }, { "epoch": 26.448275862068964, "grad_norm": 18.05221176147461, "learning_rate": 1.3154533844189018e-05, "loss": 0.1686, "step": 1534 }, { "epoch": 26.482758620689655, "grad_norm": 0.009326491504907608, "learning_rate": 1.3026819923371647e-05, "loss": 0.0008, "step": 1536 }, { "epoch": 26.517241379310345, "grad_norm": 0.04548182338476181, "learning_rate": 1.2899106002554278e-05, "loss": 0.0007, "step": 1538 }, { "epoch": 26.551724137931036, "grad_norm": 0.008745373226702213, "learning_rate": 1.277139208173691e-05, "loss": 0.0004, "step": 1540 }, { "epoch": 26.586206896551722, "grad_norm": 0.24436865746974945, "learning_rate": 1.2643678160919542e-05, "loss": 0.0021, "step": 1542 }, { "epoch": 26.620689655172413, "grad_norm": 0.01390091422945261, "learning_rate": 1.2515964240102172e-05, "loss": 0.0008, "step": 1544 }, { "epoch": 26.655172413793103, "grad_norm": 0.016228720545768738, "learning_rate": 1.2388250319284803e-05, "loss": 0.0009, "step": 1546 }, { "epoch": 26.689655172413794, "grad_norm": 1.8321493864059448, "learning_rate": 1.2260536398467433e-05, "loss": 0.0245, "step": 1548 }, { "epoch": 26.724137931034484, "grad_norm": 0.11651375144720078, "learning_rate": 1.2132822477650065e-05, "loss": 0.0012, "step": 1550 }, { "epoch": 26.75862068965517, "grad_norm": 0.26734495162963867, "learning_rate": 1.2005108556832695e-05, "loss": 0.0306, "step": 1552 }, { "epoch": 26.79310344827586, "grad_norm": 1.002111792564392, "learning_rate": 1.1877394636015327e-05, "loss": 0.0049, "step": 1554 }, { "epoch": 26.82758620689655, "grad_norm": 0.014619125053286552, "learning_rate": 1.1749680715197957e-05, "loss": 0.0012, "step": 1556 }, { "epoch": 26.862068965517242, "grad_norm": 0.18940378725528717, "learning_rate": 1.1621966794380588e-05, "loss": 0.0013, "step": 1558 }, { "epoch": 26.896551724137932, "grad_norm": 0.002744639990851283, "learning_rate": 1.1494252873563218e-05, "loss": 0.0003, "step": 1560 }, { "epoch": 26.93103448275862, "grad_norm": 0.14458243548870087, "learning_rate": 1.136653895274585e-05, "loss": 0.0017, "step": 1562 }, { "epoch": 26.96551724137931, "grad_norm": 0.44180047512054443, "learning_rate": 1.123882503192848e-05, "loss": 0.0033, "step": 1564 }, { "epoch": 27.0, "grad_norm": 1.8280788660049438, "learning_rate": 1.1111111111111112e-05, "loss": 0.0053, "step": 1566 }, { "epoch": 27.0, "eval_accuracy": 0.6666666666666666, "eval_f1_macro": 0.5719249734426743, "eval_f1_micro": 0.6666666666666666, "eval_f1_weighted": 0.6555785069889042, "eval_loss": 1.5308274030685425, "eval_precision_macro": 0.5667354028698567, "eval_precision_micro": 0.6666666666666666, "eval_precision_weighted": 0.652369328839917, "eval_recall_macro": 0.5842575187969924, "eval_recall_micro": 0.6666666666666666, "eval_recall_weighted": 0.6666666666666666, "eval_runtime": 1.9794, "eval_samples_per_second": 66.687, "eval_steps_per_second": 8.588, "step": 1566 }, { "epoch": 27.03448275862069, "grad_norm": 0.028027068823575974, "learning_rate": 1.0983397190293743e-05, "loss": 0.0007, "step": 1568 }, { "epoch": 27.06896551724138, "grad_norm": 0.017265431582927704, "learning_rate": 1.0855683269476373e-05, "loss": 0.0006, "step": 1570 }, { "epoch": 27.103448275862068, "grad_norm": 0.004702151753008366, "learning_rate": 1.0727969348659005e-05, "loss": 0.0078, "step": 1572 }, { "epoch": 27.137931034482758, "grad_norm": 1.1751693487167358, "learning_rate": 1.0600255427841635e-05, "loss": 0.0102, "step": 1574 }, { "epoch": 27.17241379310345, "grad_norm": 0.042279455810785294, "learning_rate": 1.0472541507024267e-05, "loss": 0.001, "step": 1576 }, { "epoch": 27.20689655172414, "grad_norm": 0.013697315938770771, "learning_rate": 1.0344827586206897e-05, "loss": 0.0005, "step": 1578 }, { "epoch": 27.24137931034483, "grad_norm": 0.1087651401758194, "learning_rate": 1.0217113665389528e-05, "loss": 0.0009, "step": 1580 }, { "epoch": 27.275862068965516, "grad_norm": 0.040765080600976944, "learning_rate": 1.0089399744572158e-05, "loss": 0.0015, "step": 1582 }, { "epoch": 27.310344827586206, "grad_norm": 0.15862275660037994, "learning_rate": 9.96168582375479e-06, "loss": 0.0029, "step": 1584 }, { "epoch": 27.344827586206897, "grad_norm": 0.014304181560873985, "learning_rate": 9.833971902937422e-06, "loss": 0.045, "step": 1586 }, { "epoch": 27.379310344827587, "grad_norm": 0.4837823808193207, "learning_rate": 9.706257982120052e-06, "loss": 0.003, "step": 1588 }, { "epoch": 27.413793103448278, "grad_norm": 0.00520712835714221, "learning_rate": 9.578544061302683e-06, "loss": 0.0008, "step": 1590 }, { "epoch": 27.448275862068964, "grad_norm": 0.026446396484971046, "learning_rate": 9.450830140485313e-06, "loss": 0.0005, "step": 1592 }, { "epoch": 27.482758620689655, "grad_norm": 0.021808737888932228, "learning_rate": 9.323116219667945e-06, "loss": 0.0008, "step": 1594 }, { "epoch": 27.517241379310345, "grad_norm": 1.217366337776184, "learning_rate": 9.195402298850575e-06, "loss": 0.0255, "step": 1596 }, { "epoch": 27.551724137931036, "grad_norm": 0.04630236327648163, "learning_rate": 9.067688378033207e-06, "loss": 0.0011, "step": 1598 }, { "epoch": 27.586206896551722, "grad_norm": 0.029485022649168968, "learning_rate": 8.939974457215837e-06, "loss": 0.0004, "step": 1600 }, { "epoch": 27.620689655172413, "grad_norm": 0.6011648178100586, "learning_rate": 8.812260536398467e-06, "loss": 0.0031, "step": 1602 }, { "epoch": 27.655172413793103, "grad_norm": 0.016739701852202415, "learning_rate": 8.684546615581098e-06, "loss": 0.0005, "step": 1604 }, { "epoch": 27.689655172413794, "grad_norm": 0.07989770919084549, "learning_rate": 8.55683269476373e-06, "loss": 0.0007, "step": 1606 }, { "epoch": 27.724137931034484, "grad_norm": 0.009848535992205143, "learning_rate": 8.429118773946362e-06, "loss": 0.0004, "step": 1608 }, { "epoch": 27.75862068965517, "grad_norm": 0.007213211618363857, "learning_rate": 8.301404853128992e-06, "loss": 0.0007, "step": 1610 }, { "epoch": 27.79310344827586, "grad_norm": 0.08470471948385239, "learning_rate": 8.173690932311623e-06, "loss": 0.0017, "step": 1612 }, { "epoch": 27.82758620689655, "grad_norm": 0.11105465888977051, "learning_rate": 8.045977011494253e-06, "loss": 0.001, "step": 1614 }, { "epoch": 27.862068965517242, "grad_norm": 0.38511380553245544, "learning_rate": 7.918263090676885e-06, "loss": 0.0039, "step": 1616 }, { "epoch": 27.896551724137932, "grad_norm": 0.07350783795118332, "learning_rate": 7.790549169859515e-06, "loss": 0.0011, "step": 1618 }, { "epoch": 27.93103448275862, "grad_norm": 0.0037571904249489307, "learning_rate": 7.662835249042145e-06, "loss": 0.0004, "step": 1620 }, { "epoch": 27.96551724137931, "grad_norm": 0.283407986164093, "learning_rate": 7.535121328224777e-06, "loss": 0.0016, "step": 1622 }, { "epoch": 28.0, "grad_norm": 0.025982772931456566, "learning_rate": 7.4074074074074075e-06, "loss": 0.0004, "step": 1624 }, { "epoch": 28.0, "eval_accuracy": 0.6666666666666666, "eval_f1_macro": 0.5732077187394248, "eval_f1_micro": 0.6666666666666666, "eval_f1_weighted": 0.6617322262037608, "eval_loss": 1.5639015436172485, "eval_precision_macro": 0.5684216325737007, "eval_precision_micro": 0.6666666666666666, "eval_precision_weighted": 0.667311206663744, "eval_recall_macro": 0.5867167919799499, "eval_recall_micro": 0.6666666666666666, "eval_recall_weighted": 0.6666666666666666, "eval_runtime": 2.0384, "eval_samples_per_second": 64.757, "eval_steps_per_second": 8.34, "step": 1624 }, { "epoch": 28.03448275862069, "grad_norm": 0.23507054150104523, "learning_rate": 7.279693486590039e-06, "loss": 0.003, "step": 1626 }, { "epoch": 28.06896551724138, "grad_norm": 0.0050230068154633045, "learning_rate": 7.151979565772669e-06, "loss": 0.0006, "step": 1628 }, { "epoch": 28.103448275862068, "grad_norm": 0.1437883973121643, "learning_rate": 7.024265644955301e-06, "loss": 0.0018, "step": 1630 }, { "epoch": 28.137931034482758, "grad_norm": 1.1382052898406982, "learning_rate": 6.896551724137932e-06, "loss": 0.0066, "step": 1632 }, { "epoch": 28.17241379310345, "grad_norm": 0.00969479139894247, "learning_rate": 6.768837803320563e-06, "loss": 0.0004, "step": 1634 }, { "epoch": 28.20689655172414, "grad_norm": 0.015258271247148514, "learning_rate": 6.641123882503193e-06, "loss": 0.0006, "step": 1636 }, { "epoch": 28.24137931034483, "grad_norm": 0.057105790823698044, "learning_rate": 6.513409961685823e-06, "loss": 0.0007, "step": 1638 }, { "epoch": 28.275862068965516, "grad_norm": 0.007530784234404564, "learning_rate": 6.385696040868455e-06, "loss": 0.0042, "step": 1640 }, { "epoch": 28.310344827586206, "grad_norm": 0.05503424257040024, "learning_rate": 6.257982120051086e-06, "loss": 0.0009, "step": 1642 }, { "epoch": 28.344827586206897, "grad_norm": 0.19767296314239502, "learning_rate": 6.130268199233717e-06, "loss": 0.1093, "step": 1644 }, { "epoch": 28.379310344827587, "grad_norm": 0.023296566680073738, "learning_rate": 6.0025542784163475e-06, "loss": 0.001, "step": 1646 }, { "epoch": 28.413793103448278, "grad_norm": 0.018203437328338623, "learning_rate": 5.874840357598978e-06, "loss": 0.0004, "step": 1648 }, { "epoch": 28.448275862068964, "grad_norm": 0.042694464325904846, "learning_rate": 5.747126436781609e-06, "loss": 0.0007, "step": 1650 }, { "epoch": 28.482758620689655, "grad_norm": 0.019331173971295357, "learning_rate": 5.61941251596424e-06, "loss": 0.0005, "step": 1652 }, { "epoch": 28.517241379310345, "grad_norm": 0.03333083540201187, "learning_rate": 5.491698595146872e-06, "loss": 0.0011, "step": 1654 }, { "epoch": 28.551724137931036, "grad_norm": 0.005839875899255276, "learning_rate": 5.3639846743295025e-06, "loss": 0.0006, "step": 1656 }, { "epoch": 28.586206896551722, "grad_norm": 0.040753502398729324, "learning_rate": 5.236270753512133e-06, "loss": 0.0005, "step": 1658 }, { "epoch": 28.620689655172413, "grad_norm": 0.0062615578062832355, "learning_rate": 5.108556832694764e-06, "loss": 0.0009, "step": 1660 }, { "epoch": 28.655172413793103, "grad_norm": 0.024959621950984, "learning_rate": 4.980842911877395e-06, "loss": 0.0006, "step": 1662 }, { "epoch": 28.689655172413794, "grad_norm": 0.006987102795392275, "learning_rate": 4.853128991060026e-06, "loss": 0.0007, "step": 1664 }, { "epoch": 28.724137931034484, "grad_norm": 0.005346678197383881, "learning_rate": 4.725415070242657e-06, "loss": 0.0003, "step": 1666 }, { "epoch": 28.75862068965517, "grad_norm": 0.08393517136573792, "learning_rate": 4.5977011494252875e-06, "loss": 0.0013, "step": 1668 }, { "epoch": 28.79310344827586, "grad_norm": 0.01564914733171463, "learning_rate": 4.469987228607918e-06, "loss": 0.0206, "step": 1670 }, { "epoch": 28.82758620689655, "grad_norm": 0.04356468468904495, "learning_rate": 4.342273307790549e-06, "loss": 0.0007, "step": 1672 }, { "epoch": 28.862068965517242, "grad_norm": 0.056500744074583054, "learning_rate": 4.214559386973181e-06, "loss": 0.0011, "step": 1674 }, { "epoch": 28.896551724137932, "grad_norm": 0.1394091099500656, "learning_rate": 4.086845466155812e-06, "loss": 0.0012, "step": 1676 }, { "epoch": 28.93103448275862, "grad_norm": 8.025191307067871, "learning_rate": 3.9591315453384425e-06, "loss": 0.0235, "step": 1678 }, { "epoch": 28.96551724137931, "grad_norm": 0.016247062012553215, "learning_rate": 3.8314176245210725e-06, "loss": 0.0005, "step": 1680 }, { "epoch": 29.0, "grad_norm": 0.02568941004574299, "learning_rate": 3.7037037037037037e-06, "loss": 0.0007, "step": 1682 }, { "epoch": 29.0, "eval_accuracy": 0.6742424242424242, "eval_f1_macro": 0.5835078400423941, "eval_f1_micro": 0.6742424242424242, "eval_f1_weighted": 0.6678458659891735, "eval_loss": 1.5345948934555054, "eval_precision_macro": 0.5786037923941507, "eval_precision_micro": 0.6742424242424242, "eval_precision_weighted": 0.6702666872171518, "eval_recall_macro": 0.5965382205513785, "eval_recall_micro": 0.6742424242424242, "eval_recall_weighted": 0.6742424242424242, "eval_runtime": 2.0037, "eval_samples_per_second": 65.879, "eval_steps_per_second": 8.484, "step": 1682 }, { "epoch": 29.03448275862069, "grad_norm": 0.007548394612967968, "learning_rate": 3.5759897828863346e-06, "loss": 0.0008, "step": 1684 }, { "epoch": 29.06896551724138, "grad_norm": 0.11199589818716049, "learning_rate": 3.448275862068966e-06, "loss": 0.0011, "step": 1686 }, { "epoch": 29.103448275862068, "grad_norm": 0.01992984302341938, "learning_rate": 3.3205619412515967e-06, "loss": 0.0005, "step": 1688 }, { "epoch": 29.137931034482758, "grad_norm": 0.012301865965127945, "learning_rate": 3.1928480204342275e-06, "loss": 0.0006, "step": 1690 }, { "epoch": 29.17241379310345, "grad_norm": 0.02116016112267971, "learning_rate": 3.0651340996168583e-06, "loss": 0.0097, "step": 1692 }, { "epoch": 29.20689655172414, "grad_norm": 0.020142151042819023, "learning_rate": 2.937420178799489e-06, "loss": 0.0006, "step": 1694 }, { "epoch": 29.24137931034483, "grad_norm": 0.042330626398324966, "learning_rate": 2.80970625798212e-06, "loss": 0.0009, "step": 1696 }, { "epoch": 29.275862068965516, "grad_norm": 2.6495234966278076, "learning_rate": 2.6819923371647512e-06, "loss": 0.0576, "step": 1698 }, { "epoch": 29.310344827586206, "grad_norm": 0.06072988361120224, "learning_rate": 2.554278416347382e-06, "loss": 0.0014, "step": 1700 }, { "epoch": 29.344827586206897, "grad_norm": 0.19377738237380981, "learning_rate": 2.426564495530013e-06, "loss": 0.0018, "step": 1702 }, { "epoch": 29.379310344827587, "grad_norm": 0.05790168046951294, "learning_rate": 2.2988505747126437e-06, "loss": 0.0007, "step": 1704 }, { "epoch": 29.413793103448278, "grad_norm": 0.03023909032344818, "learning_rate": 2.1711366538952746e-06, "loss": 0.04, "step": 1706 }, { "epoch": 29.448275862068964, "grad_norm": 0.039096005260944366, "learning_rate": 2.043422733077906e-06, "loss": 0.0009, "step": 1708 }, { "epoch": 29.482758620689655, "grad_norm": 1.9999027252197266, "learning_rate": 1.9157088122605362e-06, "loss": 0.0383, "step": 1710 }, { "epoch": 29.517241379310345, "grad_norm": 0.01942484639585018, "learning_rate": 1.7879948914431673e-06, "loss": 0.0004, "step": 1712 }, { "epoch": 29.551724137931036, "grad_norm": 0.019709143787622452, "learning_rate": 1.6602809706257983e-06, "loss": 0.0006, "step": 1714 }, { "epoch": 29.586206896551722, "grad_norm": 0.01487061008810997, "learning_rate": 1.5325670498084292e-06, "loss": 0.0003, "step": 1716 }, { "epoch": 29.620689655172413, "grad_norm": 0.0076087117195129395, "learning_rate": 1.40485312899106e-06, "loss": 0.0003, "step": 1718 }, { "epoch": 29.655172413793103, "grad_norm": 0.01488451100885868, "learning_rate": 1.277139208173691e-06, "loss": 0.0005, "step": 1720 }, { "epoch": 29.689655172413794, "grad_norm": 0.00445834593847394, "learning_rate": 1.1494252873563219e-06, "loss": 0.0004, "step": 1722 }, { "epoch": 29.724137931034484, "grad_norm": 0.01872408203780651, "learning_rate": 1.021711366538953e-06, "loss": 0.0005, "step": 1724 }, { "epoch": 29.75862068965517, "grad_norm": 0.010615495964884758, "learning_rate": 8.939974457215836e-07, "loss": 0.0003, "step": 1726 }, { "epoch": 29.79310344827586, "grad_norm": 0.015271478332579136, "learning_rate": 7.662835249042146e-07, "loss": 0.0005, "step": 1728 }, { "epoch": 29.82758620689655, "grad_norm": 0.2127443104982376, "learning_rate": 6.385696040868455e-07, "loss": 0.002, "step": 1730 }, { "epoch": 29.862068965517242, "grad_norm": 0.01297774724662304, "learning_rate": 5.108556832694765e-07, "loss": 0.0004, "step": 1732 }, { "epoch": 29.896551724137932, "grad_norm": 0.010099313221871853, "learning_rate": 3.831417624521073e-07, "loss": 0.0064, "step": 1734 }, { "epoch": 29.93103448275862, "grad_norm": 0.03221974894404411, "learning_rate": 2.5542784163473823e-07, "loss": 0.0005, "step": 1736 }, { "epoch": 29.96551724137931, "grad_norm": 0.11465907841920853, "learning_rate": 1.2771392081736911e-07, "loss": 0.0009, "step": 1738 }, { "epoch": 30.0, "grad_norm": 0.026466330513358116, "learning_rate": 0.0, "loss": 0.0004, "step": 1740 }, { "epoch": 30.0, "eval_accuracy": 0.6742424242424242, "eval_f1_macro": 0.5791488237085879, "eval_f1_micro": 0.6742424242424242, "eval_f1_weighted": 0.6661361118395798, "eval_loss": 1.5231688022613525, "eval_precision_macro": 0.5707257261879112, "eval_precision_micro": 0.6742424242424242, "eval_precision_weighted": 0.6628039716275012, "eval_recall_macro": 0.5917763157894738, "eval_recall_micro": 0.6742424242424242, "eval_recall_weighted": 0.6742424242424242, "eval_runtime": 2.1557, "eval_samples_per_second": 61.234, "eval_steps_per_second": 7.886, "step": 1740 }, { "epoch": 30.0, "step": 1740, "total_flos": 1.0740871074163507e+18, "train_loss": 0.5156426236089975, "train_runtime": 378.1612, "train_samples_per_second": 36.651, "train_steps_per_second": 4.601 } ], "logging_steps": 2, "max_steps": 1740, "num_input_tokens_seen": 0, "num_train_epochs": 30, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0740871074163507e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }