diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6672 @@ +{ + "best_metric": 1.0619021654129028, + "best_model_checkpoint": "square_run_second_vote/checkpoint-464", + "epoch": 30.0, + "eval_steps": 500, + "global_step": 1740, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.034482758620689655, + "grad_norm": 10.100388526916504, + "learning_rate": 1.1494252873563219e-06, + "loss": 1.9216, + "step": 2 + }, + { + "epoch": 0.06896551724137931, + "grad_norm": 15.660673141479492, + "learning_rate": 2.2988505747126437e-06, + "loss": 1.9891, + "step": 4 + }, + { + "epoch": 0.10344827586206896, + "grad_norm": 13.529168128967285, + "learning_rate": 3.448275862068966e-06, + "loss": 2.0535, + "step": 6 + }, + { + "epoch": 0.13793103448275862, + "grad_norm": 12.513371467590332, + "learning_rate": 4.5977011494252875e-06, + "loss": 2.0399, + "step": 8 + }, + { + "epoch": 0.1724137931034483, + "grad_norm": 12.855453491210938, + "learning_rate": 5.747126436781609e-06, + "loss": 1.9636, + "step": 10 + }, + { + "epoch": 0.20689655172413793, + "grad_norm": 9.419121742248535, + "learning_rate": 6.896551724137932e-06, + "loss": 1.7951, + "step": 12 + }, + { + "epoch": 0.2413793103448276, + "grad_norm": 8.58608627319336, + "learning_rate": 8.045977011494253e-06, + "loss": 1.938, + "step": 14 + }, + { + "epoch": 0.27586206896551724, + "grad_norm": 15.212312698364258, + "learning_rate": 9.195402298850575e-06, + "loss": 1.947, + "step": 16 + }, + { + "epoch": 0.3103448275862069, + "grad_norm": 10.53651237487793, + "learning_rate": 1.0344827586206897e-05, + "loss": 2.001, + "step": 18 + }, + { + "epoch": 0.3448275862068966, + "grad_norm": 10.844449043273926, + "learning_rate": 1.1494252873563218e-05, + "loss": 1.6643, + "step": 20 + }, + { + "epoch": 0.3793103448275862, + "grad_norm": 12.122403144836426, + "learning_rate": 1.2643678160919542e-05, + "loss": 1.7857, + "step": 22 + }, + { + "epoch": 0.41379310344827586, + "grad_norm": 10.027009963989258, + "learning_rate": 1.3793103448275863e-05, + "loss": 2.0627, + "step": 24 + }, + { + "epoch": 0.4482758620689655, + "grad_norm": 10.017681121826172, + "learning_rate": 1.4942528735632185e-05, + "loss": 1.8997, + "step": 26 + }, + { + "epoch": 0.4827586206896552, + "grad_norm": 11.976802825927734, + "learning_rate": 1.6091954022988507e-05, + "loss": 1.9658, + "step": 28 + }, + { + "epoch": 0.5172413793103449, + "grad_norm": 17.975006103515625, + "learning_rate": 1.7241379310344828e-05, + "loss": 2.1084, + "step": 30 + }, + { + "epoch": 0.5517241379310345, + "grad_norm": 10.179977416992188, + "learning_rate": 1.839080459770115e-05, + "loss": 1.8625, + "step": 32 + }, + { + "epoch": 0.5862068965517241, + "grad_norm": 12.881868362426758, + "learning_rate": 1.9540229885057475e-05, + "loss": 2.0918, + "step": 34 + }, + { + "epoch": 0.6206896551724138, + "grad_norm": 11.166668891906738, + "learning_rate": 2.0689655172413793e-05, + "loss": 1.7992, + "step": 36 + }, + { + "epoch": 0.6551724137931034, + "grad_norm": 9.509991645812988, + "learning_rate": 2.183908045977012e-05, + "loss": 1.7821, + "step": 38 + }, + { + "epoch": 0.6896551724137931, + "grad_norm": 9.292908668518066, + "learning_rate": 2.2988505747126437e-05, + "loss": 1.8435, + "step": 40 + }, + { + "epoch": 0.7241379310344828, + "grad_norm": 9.034067153930664, + "learning_rate": 2.413793103448276e-05, + "loss": 1.7203, + "step": 42 + }, + { + "epoch": 0.7586206896551724, + "grad_norm": 10.527994155883789, + "learning_rate": 2.5287356321839083e-05, + "loss": 2.0411, + "step": 44 + }, + { + "epoch": 0.7931034482758621, + "grad_norm": 12.680684089660645, + "learning_rate": 2.6436781609195405e-05, + "loss": 1.7733, + "step": 46 + }, + { + "epoch": 0.8275862068965517, + "grad_norm": 11.482306480407715, + "learning_rate": 2.7586206896551727e-05, + "loss": 1.9162, + "step": 48 + }, + { + "epoch": 0.8620689655172413, + "grad_norm": 9.575610160827637, + "learning_rate": 2.8735632183908045e-05, + "loss": 1.9413, + "step": 50 + }, + { + "epoch": 0.896551724137931, + "grad_norm": 9.212911605834961, + "learning_rate": 2.988505747126437e-05, + "loss": 1.6267, + "step": 52 + }, + { + "epoch": 0.9310344827586207, + "grad_norm": 10.29886245727539, + "learning_rate": 3.103448275862069e-05, + "loss": 2.0705, + "step": 54 + }, + { + "epoch": 0.9655172413793104, + "grad_norm": 10.353363990783691, + "learning_rate": 3.218390804597701e-05, + "loss": 1.7528, + "step": 56 + }, + { + "epoch": 1.0, + "grad_norm": 10.195358276367188, + "learning_rate": 3.3333333333333335e-05, + "loss": 1.8754, + "step": 58 + }, + { + "epoch": 1.0, + "eval_accuracy": 0.2803030303030303, + "eval_f1_macro": 0.13852914120300855, + "eval_f1_micro": 0.2803030303030303, + "eval_f1_weighted": 0.17218993609140762, + "eval_loss": 1.7960734367370605, + "eval_precision_macro": 0.14259740259740258, + "eval_precision_micro": 0.2803030303030303, + "eval_precision_weighted": 0.16029384756657483, + "eval_recall_macro": 0.20982142857142858, + "eval_recall_micro": 0.2803030303030303, + "eval_recall_weighted": 0.2803030303030303, + "eval_runtime": 2.0654, + "eval_samples_per_second": 63.909, + "eval_steps_per_second": 8.231, + "step": 58 + }, + { + "epoch": 1.0344827586206897, + "grad_norm": 6.763228416442871, + "learning_rate": 3.4482758620689657e-05, + "loss": 1.556, + "step": 60 + }, + { + "epoch": 1.0689655172413792, + "grad_norm": 12.80778694152832, + "learning_rate": 3.563218390804598e-05, + "loss": 1.8112, + "step": 62 + }, + { + "epoch": 1.103448275862069, + "grad_norm": 5.85200309753418, + "learning_rate": 3.67816091954023e-05, + "loss": 1.6181, + "step": 64 + }, + { + "epoch": 1.1379310344827587, + "grad_norm": 13.358109474182129, + "learning_rate": 3.793103448275862e-05, + "loss": 1.5691, + "step": 66 + }, + { + "epoch": 1.1724137931034484, + "grad_norm": 13.59086799621582, + "learning_rate": 3.908045977011495e-05, + "loss": 2.028, + "step": 68 + }, + { + "epoch": 1.206896551724138, + "grad_norm": 12.442203521728516, + "learning_rate": 4.0229885057471265e-05, + "loss": 1.8991, + "step": 70 + }, + { + "epoch": 1.2413793103448276, + "grad_norm": 10.374919891357422, + "learning_rate": 4.1379310344827587e-05, + "loss": 1.7926, + "step": 72 + }, + { + "epoch": 1.2758620689655173, + "grad_norm": 8.41528034210205, + "learning_rate": 4.252873563218391e-05, + "loss": 1.9024, + "step": 74 + }, + { + "epoch": 1.3103448275862069, + "grad_norm": 8.301165580749512, + "learning_rate": 4.367816091954024e-05, + "loss": 1.5357, + "step": 76 + }, + { + "epoch": 1.3448275862068966, + "grad_norm": 10.928062438964844, + "learning_rate": 4.482758620689655e-05, + "loss": 1.7872, + "step": 78 + }, + { + "epoch": 1.3793103448275863, + "grad_norm": 12.0917329788208, + "learning_rate": 4.597701149425287e-05, + "loss": 1.9049, + "step": 80 + }, + { + "epoch": 1.4137931034482758, + "grad_norm": 10.680750846862793, + "learning_rate": 4.7126436781609195e-05, + "loss": 1.818, + "step": 82 + }, + { + "epoch": 1.4482758620689655, + "grad_norm": 10.11230754852295, + "learning_rate": 4.827586206896552e-05, + "loss": 1.7855, + "step": 84 + }, + { + "epoch": 1.4827586206896552, + "grad_norm": 15.761420249938965, + "learning_rate": 4.9425287356321845e-05, + "loss": 1.7779, + "step": 86 + }, + { + "epoch": 1.5172413793103448, + "grad_norm": 15.814207077026367, + "learning_rate": 5.057471264367817e-05, + "loss": 1.6615, + "step": 88 + }, + { + "epoch": 1.5517241379310345, + "grad_norm": 10.473943710327148, + "learning_rate": 5.172413793103449e-05, + "loss": 1.7123, + "step": 90 + }, + { + "epoch": 1.5862068965517242, + "grad_norm": 9.527724266052246, + "learning_rate": 5.287356321839081e-05, + "loss": 1.5118, + "step": 92 + }, + { + "epoch": 1.6206896551724137, + "grad_norm": 13.386604309082031, + "learning_rate": 5.402298850574713e-05, + "loss": 2.0467, + "step": 94 + }, + { + "epoch": 1.6551724137931034, + "grad_norm": 11.433362007141113, + "learning_rate": 5.517241379310345e-05, + "loss": 1.7908, + "step": 96 + }, + { + "epoch": 1.6896551724137931, + "grad_norm": 10.231264114379883, + "learning_rate": 5.632183908045977e-05, + "loss": 2.0643, + "step": 98 + }, + { + "epoch": 1.7241379310344827, + "grad_norm": 10.50631046295166, + "learning_rate": 5.747126436781609e-05, + "loss": 1.6573, + "step": 100 + }, + { + "epoch": 1.7586206896551724, + "grad_norm": 10.044520378112793, + "learning_rate": 5.862068965517241e-05, + "loss": 1.6, + "step": 102 + }, + { + "epoch": 1.793103448275862, + "grad_norm": 11.720151901245117, + "learning_rate": 5.977011494252874e-05, + "loss": 1.9807, + "step": 104 + }, + { + "epoch": 1.8275862068965516, + "grad_norm": 8.176350593566895, + "learning_rate": 6.091954022988506e-05, + "loss": 1.772, + "step": 106 + }, + { + "epoch": 1.8620689655172413, + "grad_norm": 10.523119926452637, + "learning_rate": 6.206896551724138e-05, + "loss": 1.5979, + "step": 108 + }, + { + "epoch": 1.896551724137931, + "grad_norm": 12.917177200317383, + "learning_rate": 6.32183908045977e-05, + "loss": 1.9565, + "step": 110 + }, + { + "epoch": 1.9310344827586206, + "grad_norm": 9.43907356262207, + "learning_rate": 6.436781609195403e-05, + "loss": 1.6814, + "step": 112 + }, + { + "epoch": 1.9655172413793105, + "grad_norm": 7.691906452178955, + "learning_rate": 6.551724137931034e-05, + "loss": 1.8594, + "step": 114 + }, + { + "epoch": 2.0, + "grad_norm": 12.988773345947266, + "learning_rate": 6.666666666666667e-05, + "loss": 2.0246, + "step": 116 + }, + { + "epoch": 2.0, + "eval_accuracy": 0.3106060606060606, + "eval_f1_macro": 0.22363870840750824, + "eval_f1_micro": 0.3106060606060606, + "eval_f1_weighted": 0.2483673419087291, + "eval_loss": 2.0137689113616943, + "eval_precision_macro": 0.2557566049675684, + "eval_precision_micro": 0.3106060606060606, + "eval_precision_weighted": 0.2691759790597, + "eval_recall_macro": 0.2841635338345864, + "eval_recall_micro": 0.3106060606060606, + "eval_recall_weighted": 0.3106060606060606, + "eval_runtime": 1.9498, + "eval_samples_per_second": 67.698, + "eval_steps_per_second": 8.719, + "step": 116 + }, + { + "epoch": 2.0344827586206895, + "grad_norm": 10.840887069702148, + "learning_rate": 6.781609195402298e-05, + "loss": 1.7498, + "step": 118 + }, + { + "epoch": 2.0689655172413794, + "grad_norm": 11.36091423034668, + "learning_rate": 6.896551724137931e-05, + "loss": 1.7945, + "step": 120 + }, + { + "epoch": 2.103448275862069, + "grad_norm": 9.655730247497559, + "learning_rate": 7.011494252873564e-05, + "loss": 1.6088, + "step": 122 + }, + { + "epoch": 2.1379310344827585, + "grad_norm": 9.597773551940918, + "learning_rate": 7.126436781609196e-05, + "loss": 1.6544, + "step": 124 + }, + { + "epoch": 2.1724137931034484, + "grad_norm": 8.65636157989502, + "learning_rate": 7.241379310344828e-05, + "loss": 1.4854, + "step": 126 + }, + { + "epoch": 2.206896551724138, + "grad_norm": 8.050541877746582, + "learning_rate": 7.35632183908046e-05, + "loss": 1.3232, + "step": 128 + }, + { + "epoch": 2.2413793103448274, + "grad_norm": 11.399789810180664, + "learning_rate": 7.471264367816091e-05, + "loss": 1.6108, + "step": 130 + }, + { + "epoch": 2.2758620689655173, + "grad_norm": 10.911064147949219, + "learning_rate": 7.586206896551724e-05, + "loss": 1.3609, + "step": 132 + }, + { + "epoch": 2.310344827586207, + "grad_norm": 10.229703903198242, + "learning_rate": 7.701149425287356e-05, + "loss": 1.4551, + "step": 134 + }, + { + "epoch": 2.344827586206897, + "grad_norm": 10.077006340026855, + "learning_rate": 7.81609195402299e-05, + "loss": 1.3409, + "step": 136 + }, + { + "epoch": 2.3793103448275863, + "grad_norm": 12.368110656738281, + "learning_rate": 7.931034482758621e-05, + "loss": 1.8165, + "step": 138 + }, + { + "epoch": 2.413793103448276, + "grad_norm": 16.638439178466797, + "learning_rate": 8.045977011494253e-05, + "loss": 1.6881, + "step": 140 + }, + { + "epoch": 2.4482758620689653, + "grad_norm": 10.63574504852295, + "learning_rate": 8.160919540229886e-05, + "loss": 1.3659, + "step": 142 + }, + { + "epoch": 2.4827586206896552, + "grad_norm": 12.455951690673828, + "learning_rate": 8.275862068965517e-05, + "loss": 1.3084, + "step": 144 + }, + { + "epoch": 2.5172413793103448, + "grad_norm": 7.663853168487549, + "learning_rate": 8.39080459770115e-05, + "loss": 1.3424, + "step": 146 + }, + { + "epoch": 2.5517241379310347, + "grad_norm": 8.313244819641113, + "learning_rate": 8.505747126436782e-05, + "loss": 1.2946, + "step": 148 + }, + { + "epoch": 2.586206896551724, + "grad_norm": 9.524584770202637, + "learning_rate": 8.620689655172413e-05, + "loss": 1.1958, + "step": 150 + }, + { + "epoch": 2.6206896551724137, + "grad_norm": 12.151946067810059, + "learning_rate": 8.735632183908047e-05, + "loss": 1.6999, + "step": 152 + }, + { + "epoch": 2.655172413793103, + "grad_norm": 9.74065113067627, + "learning_rate": 8.850574712643679e-05, + "loss": 1.5755, + "step": 154 + }, + { + "epoch": 2.689655172413793, + "grad_norm": 14.421919822692871, + "learning_rate": 8.96551724137931e-05, + "loss": 1.428, + "step": 156 + }, + { + "epoch": 2.7241379310344827, + "grad_norm": 11.11561393737793, + "learning_rate": 9.080459770114943e-05, + "loss": 1.4898, + "step": 158 + }, + { + "epoch": 2.7586206896551726, + "grad_norm": 13.893324851989746, + "learning_rate": 9.195402298850575e-05, + "loss": 1.901, + "step": 160 + }, + { + "epoch": 2.793103448275862, + "grad_norm": 8.19894027709961, + "learning_rate": 9.310344827586207e-05, + "loss": 1.4159, + "step": 162 + }, + { + "epoch": 2.8275862068965516, + "grad_norm": 10.198213577270508, + "learning_rate": 9.425287356321839e-05, + "loss": 1.6808, + "step": 164 + }, + { + "epoch": 2.862068965517241, + "grad_norm": 7.003215312957764, + "learning_rate": 9.540229885057472e-05, + "loss": 1.859, + "step": 166 + }, + { + "epoch": 2.896551724137931, + "grad_norm": 8.648194313049316, + "learning_rate": 9.655172413793105e-05, + "loss": 1.5442, + "step": 168 + }, + { + "epoch": 2.9310344827586206, + "grad_norm": 9.167272567749023, + "learning_rate": 9.770114942528736e-05, + "loss": 1.6219, + "step": 170 + }, + { + "epoch": 2.9655172413793105, + "grad_norm": 13.052528381347656, + "learning_rate": 9.885057471264369e-05, + "loss": 1.637, + "step": 172 + }, + { + "epoch": 3.0, + "grad_norm": 10.723739624023438, + "learning_rate": 0.0001, + "loss": 1.6189, + "step": 174 + }, + { + "epoch": 3.0, + "eval_accuracy": 0.38636363636363635, + "eval_f1_macro": 0.244379876244283, + "eval_f1_micro": 0.38636363636363635, + "eval_f1_weighted": 0.3195200593505679, + "eval_loss": 1.5038808584213257, + "eval_precision_macro": 0.2633030990173847, + "eval_precision_micro": 0.38636363636363635, + "eval_precision_weighted": 0.33009660092993426, + "eval_recall_macro": 0.28469611528822053, + "eval_recall_micro": 0.38636363636363635, + "eval_recall_weighted": 0.38636363636363635, + "eval_runtime": 1.953, + "eval_samples_per_second": 67.59, + "eval_steps_per_second": 8.705, + "step": 174 + }, + { + "epoch": 3.0344827586206895, + "grad_norm": 7.145821571350098, + "learning_rate": 9.987228607918264e-05, + "loss": 1.3525, + "step": 176 + }, + { + "epoch": 3.0689655172413794, + "grad_norm": 8.73121166229248, + "learning_rate": 9.974457215836527e-05, + "loss": 1.3759, + "step": 178 + }, + { + "epoch": 3.103448275862069, + "grad_norm": 11.502145767211914, + "learning_rate": 9.96168582375479e-05, + "loss": 1.3372, + "step": 180 + }, + { + "epoch": 3.1379310344827585, + "grad_norm": 8.244510650634766, + "learning_rate": 9.948914431673053e-05, + "loss": 1.5859, + "step": 182 + }, + { + "epoch": 3.1724137931034484, + "grad_norm": 9.313932418823242, + "learning_rate": 9.936143039591316e-05, + "loss": 1.4753, + "step": 184 + }, + { + "epoch": 3.206896551724138, + "grad_norm": 13.148027420043945, + "learning_rate": 9.92337164750958e-05, + "loss": 1.5913, + "step": 186 + }, + { + "epoch": 3.2413793103448274, + "grad_norm": 14.681943893432617, + "learning_rate": 9.910600255427843e-05, + "loss": 1.4892, + "step": 188 + }, + { + "epoch": 3.2758620689655173, + "grad_norm": 13.758460998535156, + "learning_rate": 9.897828863346104e-05, + "loss": 1.333, + "step": 190 + }, + { + "epoch": 3.310344827586207, + "grad_norm": 11.166180610656738, + "learning_rate": 9.885057471264369e-05, + "loss": 1.5819, + "step": 192 + }, + { + "epoch": 3.344827586206897, + "grad_norm": 9.059310913085938, + "learning_rate": 9.872286079182631e-05, + "loss": 0.9963, + "step": 194 + }, + { + "epoch": 3.3793103448275863, + "grad_norm": 9.729276657104492, + "learning_rate": 9.859514687100895e-05, + "loss": 1.8236, + "step": 196 + }, + { + "epoch": 3.413793103448276, + "grad_norm": 11.96312427520752, + "learning_rate": 9.846743295019157e-05, + "loss": 1.4265, + "step": 198 + }, + { + "epoch": 3.4482758620689653, + "grad_norm": 11.733696937561035, + "learning_rate": 9.833971902937422e-05, + "loss": 1.1957, + "step": 200 + }, + { + "epoch": 3.4827586206896552, + "grad_norm": 19.109695434570312, + "learning_rate": 9.821200510855683e-05, + "loss": 2.2991, + "step": 202 + }, + { + "epoch": 3.5172413793103448, + "grad_norm": 11.54105281829834, + "learning_rate": 9.808429118773947e-05, + "loss": 1.5837, + "step": 204 + }, + { + "epoch": 3.5517241379310347, + "grad_norm": 14.453530311584473, + "learning_rate": 9.79565772669221e-05, + "loss": 1.4211, + "step": 206 + }, + { + "epoch": 3.586206896551724, + "grad_norm": 12.695812225341797, + "learning_rate": 9.782886334610473e-05, + "loss": 1.0612, + "step": 208 + }, + { + "epoch": 3.6206896551724137, + "grad_norm": 9.610103607177734, + "learning_rate": 9.770114942528736e-05, + "loss": 1.2058, + "step": 210 + }, + { + "epoch": 3.655172413793103, + "grad_norm": 12.852982521057129, + "learning_rate": 9.757343550446999e-05, + "loss": 1.735, + "step": 212 + }, + { + "epoch": 3.689655172413793, + "grad_norm": 10.032821655273438, + "learning_rate": 9.744572158365262e-05, + "loss": 1.6214, + "step": 214 + }, + { + "epoch": 3.7241379310344827, + "grad_norm": 13.28847885131836, + "learning_rate": 9.731800766283526e-05, + "loss": 1.2198, + "step": 216 + }, + { + "epoch": 3.7586206896551726, + "grad_norm": 9.55737590789795, + "learning_rate": 9.719029374201787e-05, + "loss": 1.3811, + "step": 218 + }, + { + "epoch": 3.793103448275862, + "grad_norm": 12.440496444702148, + "learning_rate": 9.706257982120052e-05, + "loss": 1.2533, + "step": 220 + }, + { + "epoch": 3.8275862068965516, + "grad_norm": 9.515503883361816, + "learning_rate": 9.693486590038314e-05, + "loss": 1.1579, + "step": 222 + }, + { + "epoch": 3.862068965517241, + "grad_norm": 14.026083946228027, + "learning_rate": 9.680715197956578e-05, + "loss": 1.7226, + "step": 224 + }, + { + "epoch": 3.896551724137931, + "grad_norm": 10.697053909301758, + "learning_rate": 9.66794380587484e-05, + "loss": 1.1127, + "step": 226 + }, + { + "epoch": 3.9310344827586206, + "grad_norm": 11.028640747070312, + "learning_rate": 9.655172413793105e-05, + "loss": 1.3266, + "step": 228 + }, + { + "epoch": 3.9655172413793105, + "grad_norm": 15.626441955566406, + "learning_rate": 9.642401021711366e-05, + "loss": 1.9315, + "step": 230 + }, + { + "epoch": 4.0, + "grad_norm": 13.396081924438477, + "learning_rate": 9.62962962962963e-05, + "loss": 1.3445, + "step": 232 + }, + { + "epoch": 4.0, + "eval_accuracy": 0.4393939393939394, + "eval_f1_macro": 0.3286744239291642, + "eval_f1_micro": 0.4393939393939394, + "eval_f1_weighted": 0.38656078267083965, + "eval_loss": 1.3982452154159546, + "eval_precision_macro": 0.31856682769726247, + "eval_precision_micro": 0.4393939393939394, + "eval_precision_weighted": 0.37300139071878197, + "eval_recall_macro": 0.3695958646616541, + "eval_recall_micro": 0.4393939393939394, + "eval_recall_weighted": 0.4393939393939394, + "eval_runtime": 1.9654, + "eval_samples_per_second": 67.161, + "eval_steps_per_second": 8.65, + "step": 232 + }, + { + "epoch": 4.0344827586206895, + "grad_norm": 9.478927612304688, + "learning_rate": 9.616858237547893e-05, + "loss": 1.3661, + "step": 234 + }, + { + "epoch": 4.068965517241379, + "grad_norm": 9.680974006652832, + "learning_rate": 9.604086845466156e-05, + "loss": 1.0935, + "step": 236 + }, + { + "epoch": 4.103448275862069, + "grad_norm": 7.175812244415283, + "learning_rate": 9.591315453384419e-05, + "loss": 1.1414, + "step": 238 + }, + { + "epoch": 4.137931034482759, + "grad_norm": 9.398384094238281, + "learning_rate": 9.578544061302682e-05, + "loss": 1.1399, + "step": 240 + }, + { + "epoch": 4.172413793103448, + "grad_norm": 6.083748817443848, + "learning_rate": 9.565772669220945e-05, + "loss": 0.8365, + "step": 242 + }, + { + "epoch": 4.206896551724138, + "grad_norm": 7.871669292449951, + "learning_rate": 9.553001277139209e-05, + "loss": 0.747, + "step": 244 + }, + { + "epoch": 4.241379310344827, + "grad_norm": 12.75833797454834, + "learning_rate": 9.540229885057472e-05, + "loss": 1.1442, + "step": 246 + }, + { + "epoch": 4.275862068965517, + "grad_norm": 9.026577949523926, + "learning_rate": 9.527458492975735e-05, + "loss": 1.1016, + "step": 248 + }, + { + "epoch": 4.310344827586207, + "grad_norm": 11.575862884521484, + "learning_rate": 9.514687100893998e-05, + "loss": 1.6752, + "step": 250 + }, + { + "epoch": 4.344827586206897, + "grad_norm": 11.908188819885254, + "learning_rate": 9.501915708812261e-05, + "loss": 1.6481, + "step": 252 + }, + { + "epoch": 4.379310344827586, + "grad_norm": 11.563265800476074, + "learning_rate": 9.489144316730524e-05, + "loss": 1.2965, + "step": 254 + }, + { + "epoch": 4.413793103448276, + "grad_norm": 8.027771949768066, + "learning_rate": 9.476372924648788e-05, + "loss": 1.1308, + "step": 256 + }, + { + "epoch": 4.448275862068965, + "grad_norm": 8.491975784301758, + "learning_rate": 9.463601532567051e-05, + "loss": 0.9251, + "step": 258 + }, + { + "epoch": 4.482758620689655, + "grad_norm": 13.935683250427246, + "learning_rate": 9.450830140485314e-05, + "loss": 1.7397, + "step": 260 + }, + { + "epoch": 4.517241379310345, + "grad_norm": 14.531791687011719, + "learning_rate": 9.438058748403577e-05, + "loss": 1.67, + "step": 262 + }, + { + "epoch": 4.551724137931035, + "grad_norm": 9.794015884399414, + "learning_rate": 9.425287356321839e-05, + "loss": 1.0594, + "step": 264 + }, + { + "epoch": 4.586206896551724, + "grad_norm": 9.404264450073242, + "learning_rate": 9.412515964240103e-05, + "loss": 1.1629, + "step": 266 + }, + { + "epoch": 4.620689655172414, + "grad_norm": 11.694725036621094, + "learning_rate": 9.399744572158365e-05, + "loss": 1.2181, + "step": 268 + }, + { + "epoch": 4.655172413793103, + "grad_norm": 12.964648246765137, + "learning_rate": 9.38697318007663e-05, + "loss": 1.2802, + "step": 270 + }, + { + "epoch": 4.689655172413794, + "grad_norm": 9.874382972717285, + "learning_rate": 9.374201787994892e-05, + "loss": 1.124, + "step": 272 + }, + { + "epoch": 4.724137931034483, + "grad_norm": 11.962569236755371, + "learning_rate": 9.361430395913156e-05, + "loss": 1.6443, + "step": 274 + }, + { + "epoch": 4.758620689655173, + "grad_norm": 9.690713882446289, + "learning_rate": 9.348659003831418e-05, + "loss": 1.6067, + "step": 276 + }, + { + "epoch": 4.793103448275862, + "grad_norm": 11.790706634521484, + "learning_rate": 9.335887611749681e-05, + "loss": 1.29, + "step": 278 + }, + { + "epoch": 4.827586206896552, + "grad_norm": 8.600451469421387, + "learning_rate": 9.323116219667944e-05, + "loss": 1.696, + "step": 280 + }, + { + "epoch": 4.862068965517241, + "grad_norm": 7.481589317321777, + "learning_rate": 9.310344827586207e-05, + "loss": 1.1751, + "step": 282 + }, + { + "epoch": 4.896551724137931, + "grad_norm": 11.036040306091309, + "learning_rate": 9.29757343550447e-05, + "loss": 0.9911, + "step": 284 + }, + { + "epoch": 4.931034482758621, + "grad_norm": 10.061944961547852, + "learning_rate": 9.284802043422734e-05, + "loss": 1.4282, + "step": 286 + }, + { + "epoch": 4.9655172413793105, + "grad_norm": 8.946602821350098, + "learning_rate": 9.272030651340997e-05, + "loss": 1.4806, + "step": 288 + }, + { + "epoch": 5.0, + "grad_norm": 8.66653823852539, + "learning_rate": 9.25925925925926e-05, + "loss": 1.3387, + "step": 290 + }, + { + "epoch": 5.0, + "eval_accuracy": 0.5757575757575758, + "eval_f1_macro": 0.4401252312339328, + "eval_f1_micro": 0.5757575757575758, + "eval_f1_weighted": 0.5265262883858709, + "eval_loss": 1.1920479536056519, + "eval_precision_macro": 0.4314786059351277, + "eval_precision_micro": 0.5757575757575758, + "eval_precision_weighted": 0.5031131056946274, + "eval_recall_macro": 0.4682644110275689, + "eval_recall_micro": 0.5757575757575758, + "eval_recall_weighted": 0.5757575757575758, + "eval_runtime": 1.976, + "eval_samples_per_second": 66.801, + "eval_steps_per_second": 8.603, + "step": 290 + }, + { + "epoch": 5.0344827586206895, + "grad_norm": 7.769369602203369, + "learning_rate": 9.246487867177522e-05, + "loss": 1.0804, + "step": 292 + }, + { + "epoch": 5.068965517241379, + "grad_norm": 6.443802356719971, + "learning_rate": 9.233716475095786e-05, + "loss": 0.7804, + "step": 294 + }, + { + "epoch": 5.103448275862069, + "grad_norm": 5.448488712310791, + "learning_rate": 9.220945083014048e-05, + "loss": 1.0786, + "step": 296 + }, + { + "epoch": 5.137931034482759, + "grad_norm": 7.302280902862549, + "learning_rate": 9.208173690932313e-05, + "loss": 1.0511, + "step": 298 + }, + { + "epoch": 5.172413793103448, + "grad_norm": 9.579891204833984, + "learning_rate": 9.195402298850575e-05, + "loss": 1.3382, + "step": 300 + }, + { + "epoch": 5.206896551724138, + "grad_norm": 8.274906158447266, + "learning_rate": 9.182630906768839e-05, + "loss": 1.0457, + "step": 302 + }, + { + "epoch": 5.241379310344827, + "grad_norm": 9.329421997070312, + "learning_rate": 9.169859514687101e-05, + "loss": 1.076, + "step": 304 + }, + { + "epoch": 5.275862068965517, + "grad_norm": 8.050252914428711, + "learning_rate": 9.157088122605364e-05, + "loss": 0.744, + "step": 306 + }, + { + "epoch": 5.310344827586207, + "grad_norm": 9.972197532653809, + "learning_rate": 9.144316730523627e-05, + "loss": 0.9465, + "step": 308 + }, + { + "epoch": 5.344827586206897, + "grad_norm": 7.221776485443115, + "learning_rate": 9.13154533844189e-05, + "loss": 1.1189, + "step": 310 + }, + { + "epoch": 5.379310344827586, + "grad_norm": 12.02575969696045, + "learning_rate": 9.118773946360154e-05, + "loss": 1.242, + "step": 312 + }, + { + "epoch": 5.413793103448276, + "grad_norm": 10.018948554992676, + "learning_rate": 9.106002554278417e-05, + "loss": 0.8253, + "step": 314 + }, + { + "epoch": 5.448275862068965, + "grad_norm": 16.198856353759766, + "learning_rate": 9.09323116219668e-05, + "loss": 2.0836, + "step": 316 + }, + { + "epoch": 5.482758620689655, + "grad_norm": 15.048312187194824, + "learning_rate": 9.080459770114943e-05, + "loss": 1.583, + "step": 318 + }, + { + "epoch": 5.517241379310345, + "grad_norm": 9.893528938293457, + "learning_rate": 9.067688378033205e-05, + "loss": 1.5501, + "step": 320 + }, + { + "epoch": 5.551724137931035, + "grad_norm": 8.719097137451172, + "learning_rate": 9.05491698595147e-05, + "loss": 1.1545, + "step": 322 + }, + { + "epoch": 5.586206896551724, + "grad_norm": 5.774980545043945, + "learning_rate": 9.042145593869731e-05, + "loss": 0.7719, + "step": 324 + }, + { + "epoch": 5.620689655172414, + "grad_norm": 6.541261672973633, + "learning_rate": 9.029374201787996e-05, + "loss": 0.8762, + "step": 326 + }, + { + "epoch": 5.655172413793103, + "grad_norm": 8.278448104858398, + "learning_rate": 9.016602809706258e-05, + "loss": 0.9054, + "step": 328 + }, + { + "epoch": 5.689655172413794, + "grad_norm": 8.492527961730957, + "learning_rate": 9.003831417624522e-05, + "loss": 0.8369, + "step": 330 + }, + { + "epoch": 5.724137931034483, + "grad_norm": 10.219111442565918, + "learning_rate": 8.991060025542784e-05, + "loss": 1.2223, + "step": 332 + }, + { + "epoch": 5.758620689655173, + "grad_norm": 8.872493743896484, + "learning_rate": 8.978288633461047e-05, + "loss": 1.3745, + "step": 334 + }, + { + "epoch": 5.793103448275862, + "grad_norm": 7.07079553604126, + "learning_rate": 8.96551724137931e-05, + "loss": 0.8249, + "step": 336 + }, + { + "epoch": 5.827586206896552, + "grad_norm": 11.903797149658203, + "learning_rate": 8.952745849297573e-05, + "loss": 1.4273, + "step": 338 + }, + { + "epoch": 5.862068965517241, + "grad_norm": 7.793108940124512, + "learning_rate": 8.939974457215837e-05, + "loss": 0.8989, + "step": 340 + }, + { + "epoch": 5.896551724137931, + "grad_norm": 7.905730247497559, + "learning_rate": 8.9272030651341e-05, + "loss": 1.1299, + "step": 342 + }, + { + "epoch": 5.931034482758621, + "grad_norm": 8.985093116760254, + "learning_rate": 8.914431673052363e-05, + "loss": 1.356, + "step": 344 + }, + { + "epoch": 5.9655172413793105, + "grad_norm": 18.639375686645508, + "learning_rate": 8.901660280970626e-05, + "loss": 1.4211, + "step": 346 + }, + { + "epoch": 6.0, + "grad_norm": 12.543028831481934, + "learning_rate": 8.888888888888889e-05, + "loss": 1.1664, + "step": 348 + }, + { + "epoch": 6.0, + "eval_accuracy": 0.5075757575757576, + "eval_f1_macro": 0.41789107824630456, + "eval_f1_micro": 0.5075757575757576, + "eval_f1_weighted": 0.49881339629823973, + "eval_loss": 1.177772879600525, + "eval_precision_macro": 0.506774193548387, + "eval_precision_micro": 0.5075757575757576, + "eval_precision_weighted": 0.5861559139784946, + "eval_recall_macro": 0.43952067669172934, + "eval_recall_micro": 0.5075757575757576, + "eval_recall_weighted": 0.5075757575757576, + "eval_runtime": 1.9845, + "eval_samples_per_second": 66.515, + "eval_steps_per_second": 8.566, + "step": 348 + }, + { + "epoch": 6.0344827586206895, + "grad_norm": 10.878969192504883, + "learning_rate": 8.876117496807152e-05, + "loss": 1.2538, + "step": 350 + }, + { + "epoch": 6.068965517241379, + "grad_norm": 16.26839828491211, + "learning_rate": 8.863346104725416e-05, + "loss": 2.0077, + "step": 352 + }, + { + "epoch": 6.103448275862069, + "grad_norm": 4.248413562774658, + "learning_rate": 8.850574712643679e-05, + "loss": 0.6126, + "step": 354 + }, + { + "epoch": 6.137931034482759, + "grad_norm": 8.243701934814453, + "learning_rate": 8.837803320561942e-05, + "loss": 0.9125, + "step": 356 + }, + { + "epoch": 6.172413793103448, + "grad_norm": 10.138152122497559, + "learning_rate": 8.825031928480205e-05, + "loss": 0.7693, + "step": 358 + }, + { + "epoch": 6.206896551724138, + "grad_norm": 9.491286277770996, + "learning_rate": 8.812260536398468e-05, + "loss": 1.1705, + "step": 360 + }, + { + "epoch": 6.241379310344827, + "grad_norm": 10.823892593383789, + "learning_rate": 8.79948914431673e-05, + "loss": 1.4309, + "step": 362 + }, + { + "epoch": 6.275862068965517, + "grad_norm": 9.899202346801758, + "learning_rate": 8.786717752234995e-05, + "loss": 1.0038, + "step": 364 + }, + { + "epoch": 6.310344827586207, + "grad_norm": 8.473255157470703, + "learning_rate": 8.773946360153256e-05, + "loss": 1.1241, + "step": 366 + }, + { + "epoch": 6.344827586206897, + "grad_norm": 8.142688751220703, + "learning_rate": 8.761174968071521e-05, + "loss": 0.9774, + "step": 368 + }, + { + "epoch": 6.379310344827586, + "grad_norm": 8.222920417785645, + "learning_rate": 8.748403575989783e-05, + "loss": 1.105, + "step": 370 + }, + { + "epoch": 6.413793103448276, + "grad_norm": 8.64330768585205, + "learning_rate": 8.735632183908047e-05, + "loss": 0.7421, + "step": 372 + }, + { + "epoch": 6.448275862068965, + "grad_norm": 8.659346580505371, + "learning_rate": 8.722860791826309e-05, + "loss": 0.8458, + "step": 374 + }, + { + "epoch": 6.482758620689655, + "grad_norm": 7.194849967956543, + "learning_rate": 8.710089399744572e-05, + "loss": 0.6802, + "step": 376 + }, + { + "epoch": 6.517241379310345, + "grad_norm": 8.775270462036133, + "learning_rate": 8.697318007662835e-05, + "loss": 1.1627, + "step": 378 + }, + { + "epoch": 6.551724137931035, + "grad_norm": 9.507420539855957, + "learning_rate": 8.684546615581099e-05, + "loss": 0.9029, + "step": 380 + }, + { + "epoch": 6.586206896551724, + "grad_norm": 9.500436782836914, + "learning_rate": 8.671775223499362e-05, + "loss": 1.1927, + "step": 382 + }, + { + "epoch": 6.620689655172414, + "grad_norm": 2.656595230102539, + "learning_rate": 8.659003831417625e-05, + "loss": 0.8011, + "step": 384 + }, + { + "epoch": 6.655172413793103, + "grad_norm": 11.397368431091309, + "learning_rate": 8.646232439335888e-05, + "loss": 1.3364, + "step": 386 + }, + { + "epoch": 6.689655172413794, + "grad_norm": 9.544795036315918, + "learning_rate": 8.633461047254151e-05, + "loss": 0.9104, + "step": 388 + }, + { + "epoch": 6.724137931034483, + "grad_norm": 9.221898078918457, + "learning_rate": 8.620689655172413e-05, + "loss": 1.3124, + "step": 390 + }, + { + "epoch": 6.758620689655173, + "grad_norm": 8.60506534576416, + "learning_rate": 8.607918263090678e-05, + "loss": 0.8603, + "step": 392 + }, + { + "epoch": 6.793103448275862, + "grad_norm": 3.3715968132019043, + "learning_rate": 8.59514687100894e-05, + "loss": 0.8369, + "step": 394 + }, + { + "epoch": 6.827586206896552, + "grad_norm": 7.430781841278076, + "learning_rate": 8.582375478927204e-05, + "loss": 0.7418, + "step": 396 + }, + { + "epoch": 6.862068965517241, + "grad_norm": 11.155794143676758, + "learning_rate": 8.569604086845466e-05, + "loss": 1.2551, + "step": 398 + }, + { + "epoch": 6.896551724137931, + "grad_norm": 12.489243507385254, + "learning_rate": 8.55683269476373e-05, + "loss": 1.3012, + "step": 400 + }, + { + "epoch": 6.931034482758621, + "grad_norm": 10.465293884277344, + "learning_rate": 8.544061302681992e-05, + "loss": 1.122, + "step": 402 + }, + { + "epoch": 6.9655172413793105, + "grad_norm": 7.462335109710693, + "learning_rate": 8.531289910600255e-05, + "loss": 1.2381, + "step": 404 + }, + { + "epoch": 7.0, + "grad_norm": 8.508047103881836, + "learning_rate": 8.518518518518518e-05, + "loss": 1.1622, + "step": 406 + }, + { + "epoch": 7.0, + "eval_accuracy": 0.5378787878787878, + "eval_f1_macro": 0.45176980273325784, + "eval_f1_micro": 0.5378787878787878, + "eval_f1_weighted": 0.52507696437929, + "eval_loss": 1.1722992658615112, + "eval_precision_macro": 0.45142096229052747, + "eval_precision_micro": 0.5378787878787878, + "eval_precision_weighted": 0.5525594574507617, + "eval_recall_macro": 0.4867324561403509, + "eval_recall_micro": 0.5378787878787878, + "eval_recall_weighted": 0.5378787878787878, + "eval_runtime": 1.9976, + "eval_samples_per_second": 66.081, + "eval_steps_per_second": 8.51, + "step": 406 + }, + { + "epoch": 7.0344827586206895, + "grad_norm": 9.81274700164795, + "learning_rate": 8.505747126436782e-05, + "loss": 0.8411, + "step": 408 + }, + { + "epoch": 7.068965517241379, + "grad_norm": 9.398898124694824, + "learning_rate": 8.492975734355045e-05, + "loss": 1.0148, + "step": 410 + }, + { + "epoch": 7.103448275862069, + "grad_norm": 8.494013786315918, + "learning_rate": 8.480204342273308e-05, + "loss": 0.5251, + "step": 412 + }, + { + "epoch": 7.137931034482759, + "grad_norm": 8.39294147491455, + "learning_rate": 8.467432950191571e-05, + "loss": 0.6457, + "step": 414 + }, + { + "epoch": 7.172413793103448, + "grad_norm": 12.938483238220215, + "learning_rate": 8.454661558109834e-05, + "loss": 1.0682, + "step": 416 + }, + { + "epoch": 7.206896551724138, + "grad_norm": 8.100533485412598, + "learning_rate": 8.441890166028097e-05, + "loss": 1.0759, + "step": 418 + }, + { + "epoch": 7.241379310344827, + "grad_norm": 9.130753517150879, + "learning_rate": 8.42911877394636e-05, + "loss": 0.665, + "step": 420 + }, + { + "epoch": 7.275862068965517, + "grad_norm": 7.9572906494140625, + "learning_rate": 8.416347381864624e-05, + "loss": 0.5862, + "step": 422 + }, + { + "epoch": 7.310344827586207, + "grad_norm": 7.2201104164123535, + "learning_rate": 8.403575989782887e-05, + "loss": 0.8631, + "step": 424 + }, + { + "epoch": 7.344827586206897, + "grad_norm": 9.218771934509277, + "learning_rate": 8.39080459770115e-05, + "loss": 0.9256, + "step": 426 + }, + { + "epoch": 7.379310344827586, + "grad_norm": 10.637073516845703, + "learning_rate": 8.378033205619413e-05, + "loss": 0.5491, + "step": 428 + }, + { + "epoch": 7.413793103448276, + "grad_norm": 6.346437931060791, + "learning_rate": 8.365261813537676e-05, + "loss": 0.7555, + "step": 430 + }, + { + "epoch": 7.448275862068965, + "grad_norm": 5.5799384117126465, + "learning_rate": 8.35249042145594e-05, + "loss": 0.3958, + "step": 432 + }, + { + "epoch": 7.482758620689655, + "grad_norm": 11.625855445861816, + "learning_rate": 8.339719029374203e-05, + "loss": 0.7716, + "step": 434 + }, + { + "epoch": 7.517241379310345, + "grad_norm": 11.031689643859863, + "learning_rate": 8.326947637292465e-05, + "loss": 0.946, + "step": 436 + }, + { + "epoch": 7.551724137931035, + "grad_norm": 14.949385643005371, + "learning_rate": 8.314176245210729e-05, + "loss": 1.448, + "step": 438 + }, + { + "epoch": 7.586206896551724, + "grad_norm": 9.996256828308105, + "learning_rate": 8.301404853128991e-05, + "loss": 1.0621, + "step": 440 + }, + { + "epoch": 7.620689655172414, + "grad_norm": 12.930780410766602, + "learning_rate": 8.288633461047255e-05, + "loss": 1.5245, + "step": 442 + }, + { + "epoch": 7.655172413793103, + "grad_norm": 9.777554512023926, + "learning_rate": 8.275862068965517e-05, + "loss": 0.7166, + "step": 444 + }, + { + "epoch": 7.689655172413794, + "grad_norm": 15.452765464782715, + "learning_rate": 8.263090676883782e-05, + "loss": 0.81, + "step": 446 + }, + { + "epoch": 7.724137931034483, + "grad_norm": 5.021923542022705, + "learning_rate": 8.250319284802044e-05, + "loss": 0.4846, + "step": 448 + }, + { + "epoch": 7.758620689655173, + "grad_norm": 10.42216682434082, + "learning_rate": 8.237547892720307e-05, + "loss": 0.8314, + "step": 450 + }, + { + "epoch": 7.793103448275862, + "grad_norm": 10.091817855834961, + "learning_rate": 8.22477650063857e-05, + "loss": 0.7462, + "step": 452 + }, + { + "epoch": 7.827586206896552, + "grad_norm": 10.696878433227539, + "learning_rate": 8.212005108556833e-05, + "loss": 1.1465, + "step": 454 + }, + { + "epoch": 7.862068965517241, + "grad_norm": 13.578857421875, + "learning_rate": 8.199233716475096e-05, + "loss": 1.7135, + "step": 456 + }, + { + "epoch": 7.896551724137931, + "grad_norm": 8.861135482788086, + "learning_rate": 8.18646232439336e-05, + "loss": 0.9397, + "step": 458 + }, + { + "epoch": 7.931034482758621, + "grad_norm": 8.846278190612793, + "learning_rate": 8.173690932311623e-05, + "loss": 0.932, + "step": 460 + }, + { + "epoch": 7.9655172413793105, + "grad_norm": 13.574858665466309, + "learning_rate": 8.160919540229886e-05, + "loss": 1.6147, + "step": 462 + }, + { + "epoch": 8.0, + "grad_norm": 11.378049850463867, + "learning_rate": 8.148148148148148e-05, + "loss": 0.9827, + "step": 464 + }, + { + "epoch": 8.0, + "eval_accuracy": 0.6212121212121212, + "eval_f1_macro": 0.5083652151343768, + "eval_f1_micro": 0.6212121212121212, + "eval_f1_weighted": 0.6074328355724158, + "eval_loss": 1.0619021654129028, + "eval_precision_macro": 0.5036885527157723, + "eval_precision_micro": 0.6212121212121212, + "eval_precision_weighted": 0.61396284194398, + "eval_recall_macro": 0.5344924812030075, + "eval_recall_micro": 0.6212121212121212, + "eval_recall_weighted": 0.6212121212121212, + "eval_runtime": 1.9917, + "eval_samples_per_second": 66.275, + "eval_steps_per_second": 8.535, + "step": 464 + }, + { + "epoch": 8.03448275862069, + "grad_norm": 5.481504440307617, + "learning_rate": 8.135376756066412e-05, + "loss": 0.4917, + "step": 466 + }, + { + "epoch": 8.068965517241379, + "grad_norm": 7.311232566833496, + "learning_rate": 8.122605363984674e-05, + "loss": 0.7269, + "step": 468 + }, + { + "epoch": 8.10344827586207, + "grad_norm": 9.246880531311035, + "learning_rate": 8.109833971902938e-05, + "loss": 0.8737, + "step": 470 + }, + { + "epoch": 8.137931034482758, + "grad_norm": 8.400725364685059, + "learning_rate": 8.0970625798212e-05, + "loss": 0.5171, + "step": 472 + }, + { + "epoch": 8.172413793103448, + "grad_norm": 7.584453105926514, + "learning_rate": 8.084291187739465e-05, + "loss": 0.4877, + "step": 474 + }, + { + "epoch": 8.206896551724139, + "grad_norm": 3.6026341915130615, + "learning_rate": 8.071519795657727e-05, + "loss": 0.3995, + "step": 476 + }, + { + "epoch": 8.241379310344827, + "grad_norm": 11.562987327575684, + "learning_rate": 8.05874840357599e-05, + "loss": 0.7385, + "step": 478 + }, + { + "epoch": 8.275862068965518, + "grad_norm": 6.191165447235107, + "learning_rate": 8.045977011494253e-05, + "loss": 0.5029, + "step": 480 + }, + { + "epoch": 8.310344827586206, + "grad_norm": 6.5581583976745605, + "learning_rate": 8.033205619412516e-05, + "loss": 0.4527, + "step": 482 + }, + { + "epoch": 8.344827586206897, + "grad_norm": 5.58575963973999, + "learning_rate": 8.020434227330779e-05, + "loss": 0.505, + "step": 484 + }, + { + "epoch": 8.379310344827585, + "grad_norm": 7.359447956085205, + "learning_rate": 8.007662835249042e-05, + "loss": 0.3806, + "step": 486 + }, + { + "epoch": 8.413793103448276, + "grad_norm": 9.096012115478516, + "learning_rate": 7.994891443167306e-05, + "loss": 0.7836, + "step": 488 + }, + { + "epoch": 8.448275862068966, + "grad_norm": 5.758499622344971, + "learning_rate": 7.982120051085569e-05, + "loss": 0.6284, + "step": 490 + }, + { + "epoch": 8.482758620689655, + "grad_norm": 9.046234130859375, + "learning_rate": 7.969348659003832e-05, + "loss": 0.8215, + "step": 492 + }, + { + "epoch": 8.517241379310345, + "grad_norm": 10.501338958740234, + "learning_rate": 7.956577266922095e-05, + "loss": 0.7549, + "step": 494 + }, + { + "epoch": 8.551724137931034, + "grad_norm": 12.925354957580566, + "learning_rate": 7.943805874840358e-05, + "loss": 0.9301, + "step": 496 + }, + { + "epoch": 8.586206896551724, + "grad_norm": 9.080878257751465, + "learning_rate": 7.931034482758621e-05, + "loss": 0.9217, + "step": 498 + }, + { + "epoch": 8.620689655172415, + "grad_norm": 11.387174606323242, + "learning_rate": 7.918263090676885e-05, + "loss": 1.1331, + "step": 500 + }, + { + "epoch": 8.655172413793103, + "grad_norm": 4.864457130432129, + "learning_rate": 7.905491698595148e-05, + "loss": 0.3399, + "step": 502 + }, + { + "epoch": 8.689655172413794, + "grad_norm": 9.243449211120605, + "learning_rate": 7.892720306513411e-05, + "loss": 0.8055, + "step": 504 + }, + { + "epoch": 8.724137931034482, + "grad_norm": 13.595425605773926, + "learning_rate": 7.879948914431673e-05, + "loss": 1.2361, + "step": 506 + }, + { + "epoch": 8.758620689655173, + "grad_norm": 9.03889274597168, + "learning_rate": 7.867177522349937e-05, + "loss": 1.2648, + "step": 508 + }, + { + "epoch": 8.793103448275861, + "grad_norm": 10.691272735595703, + "learning_rate": 7.854406130268199e-05, + "loss": 0.8417, + "step": 510 + }, + { + "epoch": 8.827586206896552, + "grad_norm": 15.45632266998291, + "learning_rate": 7.841634738186464e-05, + "loss": 1.1186, + "step": 512 + }, + { + "epoch": 8.862068965517242, + "grad_norm": 10.35450267791748, + "learning_rate": 7.828863346104725e-05, + "loss": 1.0734, + "step": 514 + }, + { + "epoch": 8.89655172413793, + "grad_norm": 5.785059452056885, + "learning_rate": 7.81609195402299e-05, + "loss": 0.5601, + "step": 516 + }, + { + "epoch": 8.931034482758621, + "grad_norm": 8.064878463745117, + "learning_rate": 7.803320561941252e-05, + "loss": 0.8273, + "step": 518 + }, + { + "epoch": 8.96551724137931, + "grad_norm": 9.697986602783203, + "learning_rate": 7.790549169859515e-05, + "loss": 0.7774, + "step": 520 + }, + { + "epoch": 9.0, + "grad_norm": 14.628928184509277, + "learning_rate": 7.777777777777778e-05, + "loss": 1.3416, + "step": 522 + }, + { + "epoch": 9.0, + "eval_accuracy": 0.5, + "eval_f1_macro": 0.3996626928502454, + "eval_f1_micro": 0.5, + "eval_f1_weighted": 0.4689747329928523, + "eval_loss": 1.3995325565338135, + "eval_precision_macro": 0.4217794875689612, + "eval_precision_micro": 0.5, + "eval_precision_weighted": 0.5023755373635757, + "eval_recall_macro": 0.45092418546365914, + "eval_recall_micro": 0.5, + "eval_recall_weighted": 0.5, + "eval_runtime": 1.9802, + "eval_samples_per_second": 66.659, + "eval_steps_per_second": 8.585, + "step": 522 + }, + { + "epoch": 9.03448275862069, + "grad_norm": 8.5454683303833, + "learning_rate": 7.765006385696041e-05, + "loss": 0.7221, + "step": 524 + }, + { + "epoch": 9.068965517241379, + "grad_norm": 12.215627670288086, + "learning_rate": 7.752234993614304e-05, + "loss": 1.1389, + "step": 526 + }, + { + "epoch": 9.10344827586207, + "grad_norm": 7.772884368896484, + "learning_rate": 7.739463601532568e-05, + "loss": 0.5547, + "step": 528 + }, + { + "epoch": 9.137931034482758, + "grad_norm": 6.657981872558594, + "learning_rate": 7.726692209450831e-05, + "loss": 0.4854, + "step": 530 + }, + { + "epoch": 9.172413793103448, + "grad_norm": 4.549035549163818, + "learning_rate": 7.713920817369094e-05, + "loss": 0.4071, + "step": 532 + }, + { + "epoch": 9.206896551724139, + "grad_norm": 10.139058113098145, + "learning_rate": 7.701149425287356e-05, + "loss": 0.8507, + "step": 534 + }, + { + "epoch": 9.241379310344827, + "grad_norm": 9.346701622009277, + "learning_rate": 7.68837803320562e-05, + "loss": 0.5146, + "step": 536 + }, + { + "epoch": 9.275862068965518, + "grad_norm": 8.409459114074707, + "learning_rate": 7.675606641123882e-05, + "loss": 0.7852, + "step": 538 + }, + { + "epoch": 9.310344827586206, + "grad_norm": 10.624256134033203, + "learning_rate": 7.662835249042147e-05, + "loss": 0.7101, + "step": 540 + }, + { + "epoch": 9.344827586206897, + "grad_norm": 10.048233032226562, + "learning_rate": 7.650063856960408e-05, + "loss": 0.9228, + "step": 542 + }, + { + "epoch": 9.379310344827585, + "grad_norm": 7.511538982391357, + "learning_rate": 7.637292464878673e-05, + "loss": 0.4572, + "step": 544 + }, + { + "epoch": 9.413793103448276, + "grad_norm": 5.808769226074219, + "learning_rate": 7.624521072796935e-05, + "loss": 0.3693, + "step": 546 + }, + { + "epoch": 9.448275862068966, + "grad_norm": 7.7615156173706055, + "learning_rate": 7.611749680715198e-05, + "loss": 0.5247, + "step": 548 + }, + { + "epoch": 9.482758620689655, + "grad_norm": 13.348309516906738, + "learning_rate": 7.598978288633461e-05, + "loss": 1.0443, + "step": 550 + }, + { + "epoch": 9.517241379310345, + "grad_norm": 9.64974594116211, + "learning_rate": 7.586206896551724e-05, + "loss": 0.563, + "step": 552 + }, + { + "epoch": 9.551724137931034, + "grad_norm": 11.617941856384277, + "learning_rate": 7.573435504469987e-05, + "loss": 1.021, + "step": 554 + }, + { + "epoch": 9.586206896551724, + "grad_norm": 8.760254859924316, + "learning_rate": 7.56066411238825e-05, + "loss": 0.7207, + "step": 556 + }, + { + "epoch": 9.620689655172415, + "grad_norm": 9.84865951538086, + "learning_rate": 7.547892720306514e-05, + "loss": 0.4963, + "step": 558 + }, + { + "epoch": 9.655172413793103, + "grad_norm": 9.160222053527832, + "learning_rate": 7.535121328224777e-05, + "loss": 0.8536, + "step": 560 + }, + { + "epoch": 9.689655172413794, + "grad_norm": 6.880321502685547, + "learning_rate": 7.522349936143039e-05, + "loss": 1.1081, + "step": 562 + }, + { + "epoch": 9.724137931034482, + "grad_norm": 7.537998676300049, + "learning_rate": 7.509578544061303e-05, + "loss": 0.5926, + "step": 564 + }, + { + "epoch": 9.758620689655173, + "grad_norm": 10.582006454467773, + "learning_rate": 7.496807151979565e-05, + "loss": 0.9216, + "step": 566 + }, + { + "epoch": 9.793103448275861, + "grad_norm": 8.612391471862793, + "learning_rate": 7.48403575989783e-05, + "loss": 0.4138, + "step": 568 + }, + { + "epoch": 9.827586206896552, + "grad_norm": 9.592196464538574, + "learning_rate": 7.471264367816091e-05, + "loss": 0.7652, + "step": 570 + }, + { + "epoch": 9.862068965517242, + "grad_norm": 7.557244300842285, + "learning_rate": 7.458492975734356e-05, + "loss": 0.4475, + "step": 572 + }, + { + "epoch": 9.89655172413793, + "grad_norm": 12.449685096740723, + "learning_rate": 7.445721583652618e-05, + "loss": 0.7721, + "step": 574 + }, + { + "epoch": 9.931034482758621, + "grad_norm": 7.203516960144043, + "learning_rate": 7.432950191570882e-05, + "loss": 1.0207, + "step": 576 + }, + { + "epoch": 9.96551724137931, + "grad_norm": 4.533941745758057, + "learning_rate": 7.420178799489144e-05, + "loss": 0.5956, + "step": 578 + }, + { + "epoch": 10.0, + "grad_norm": 21.596515655517578, + "learning_rate": 7.407407407407407e-05, + "loss": 0.758, + "step": 580 + }, + { + "epoch": 10.0, + "eval_accuracy": 0.5984848484848485, + "eval_f1_macro": 0.5065544106205532, + "eval_f1_micro": 0.5984848484848485, + "eval_f1_weighted": 0.5835752856194464, + "eval_loss": 1.1693105697631836, + "eval_precision_macro": 0.5261672850958565, + "eval_precision_micro": 0.5984848484848485, + "eval_precision_weighted": 0.6030692476715204, + "eval_recall_macro": 0.5278665413533835, + "eval_recall_micro": 0.5984848484848485, + "eval_recall_weighted": 0.5984848484848485, + "eval_runtime": 2.0002, + "eval_samples_per_second": 65.992, + "eval_steps_per_second": 8.499, + "step": 580 + }, + { + "epoch": 10.03448275862069, + "grad_norm": 5.144809246063232, + "learning_rate": 7.39463601532567e-05, + "loss": 0.2258, + "step": 582 + }, + { + "epoch": 10.068965517241379, + "grad_norm": 8.242935180664062, + "learning_rate": 7.381864623243934e-05, + "loss": 0.45, + "step": 584 + }, + { + "epoch": 10.10344827586207, + "grad_norm": 3.4672491550445557, + "learning_rate": 7.369093231162197e-05, + "loss": 0.4987, + "step": 586 + }, + { + "epoch": 10.137931034482758, + "grad_norm": 1.9440521001815796, + "learning_rate": 7.35632183908046e-05, + "loss": 0.2707, + "step": 588 + }, + { + "epoch": 10.172413793103448, + "grad_norm": 11.417020797729492, + "learning_rate": 7.343550446998723e-05, + "loss": 0.6887, + "step": 590 + }, + { + "epoch": 10.206896551724139, + "grad_norm": 6.448490142822266, + "learning_rate": 7.330779054916986e-05, + "loss": 0.3351, + "step": 592 + }, + { + "epoch": 10.241379310344827, + "grad_norm": 21.801414489746094, + "learning_rate": 7.31800766283525e-05, + "loss": 0.8969, + "step": 594 + }, + { + "epoch": 10.275862068965518, + "grad_norm": 3.7770888805389404, + "learning_rate": 7.305236270753513e-05, + "loss": 0.3175, + "step": 596 + }, + { + "epoch": 10.310344827586206, + "grad_norm": 12.79362678527832, + "learning_rate": 7.292464878671776e-05, + "loss": 0.756, + "step": 598 + }, + { + "epoch": 10.344827586206897, + "grad_norm": 12.172070503234863, + "learning_rate": 7.279693486590039e-05, + "loss": 0.6943, + "step": 600 + }, + { + "epoch": 10.379310344827585, + "grad_norm": 11.363933563232422, + "learning_rate": 7.266922094508302e-05, + "loss": 0.7048, + "step": 602 + }, + { + "epoch": 10.413793103448276, + "grad_norm": 7.672904968261719, + "learning_rate": 7.254150702426565e-05, + "loss": 0.2959, + "step": 604 + }, + { + "epoch": 10.448275862068966, + "grad_norm": 9.661025047302246, + "learning_rate": 7.241379310344828e-05, + "loss": 0.4829, + "step": 606 + }, + { + "epoch": 10.482758620689655, + "grad_norm": 8.507035255432129, + "learning_rate": 7.22860791826309e-05, + "loss": 0.2263, + "step": 608 + }, + { + "epoch": 10.517241379310345, + "grad_norm": 6.187190055847168, + "learning_rate": 7.215836526181355e-05, + "loss": 0.5414, + "step": 610 + }, + { + "epoch": 10.551724137931034, + "grad_norm": 11.629401206970215, + "learning_rate": 7.203065134099617e-05, + "loss": 0.7474, + "step": 612 + }, + { + "epoch": 10.586206896551724, + "grad_norm": 7.820981025695801, + "learning_rate": 7.190293742017881e-05, + "loss": 0.4997, + "step": 614 + }, + { + "epoch": 10.620689655172415, + "grad_norm": 14.111104011535645, + "learning_rate": 7.177522349936143e-05, + "loss": 0.4296, + "step": 616 + }, + { + "epoch": 10.655172413793103, + "grad_norm": 10.963750839233398, + "learning_rate": 7.164750957854408e-05, + "loss": 0.4871, + "step": 618 + }, + { + "epoch": 10.689655172413794, + "grad_norm": 17.322771072387695, + "learning_rate": 7.151979565772669e-05, + "loss": 0.9553, + "step": 620 + }, + { + "epoch": 10.724137931034482, + "grad_norm": 9.986748695373535, + "learning_rate": 7.139208173690932e-05, + "loss": 0.5509, + "step": 622 + }, + { + "epoch": 10.758620689655173, + "grad_norm": 9.360352516174316, + "learning_rate": 7.126436781609196e-05, + "loss": 0.6703, + "step": 624 + }, + { + "epoch": 10.793103448275861, + "grad_norm": 3.3035664558410645, + "learning_rate": 7.113665389527459e-05, + "loss": 0.4547, + "step": 626 + }, + { + "epoch": 10.827586206896552, + "grad_norm": 8.55704116821289, + "learning_rate": 7.100893997445722e-05, + "loss": 0.7755, + "step": 628 + }, + { + "epoch": 10.862068965517242, + "grad_norm": 12.647590637207031, + "learning_rate": 7.088122605363985e-05, + "loss": 0.8421, + "step": 630 + }, + { + "epoch": 10.89655172413793, + "grad_norm": 4.907244682312012, + "learning_rate": 7.075351213282248e-05, + "loss": 0.3781, + "step": 632 + }, + { + "epoch": 10.931034482758621, + "grad_norm": 10.101003646850586, + "learning_rate": 7.062579821200511e-05, + "loss": 0.6021, + "step": 634 + }, + { + "epoch": 10.96551724137931, + "grad_norm": 15.203731536865234, + "learning_rate": 7.049808429118773e-05, + "loss": 0.9916, + "step": 636 + }, + { + "epoch": 11.0, + "grad_norm": 20.572900772094727, + "learning_rate": 7.037037037037038e-05, + "loss": 0.7758, + "step": 638 + }, + { + "epoch": 11.0, + "eval_accuracy": 0.6515151515151515, + "eval_f1_macro": 0.5490749541247198, + "eval_f1_micro": 0.6515151515151515, + "eval_f1_weighted": 0.6319976363055722, + "eval_loss": 1.0799717903137207, + "eval_precision_macro": 0.5729282057650646, + "eval_precision_micro": 0.6515151515151515, + "eval_precision_weighted": 0.6500887606871381, + "eval_recall_macro": 0.5710213032581454, + "eval_recall_micro": 0.6515151515151515, + "eval_recall_weighted": 0.6515151515151515, + "eval_runtime": 1.9879, + "eval_samples_per_second": 66.401, + "eval_steps_per_second": 8.552, + "step": 638 + }, + { + "epoch": 11.03448275862069, + "grad_norm": 6.214511871337891, + "learning_rate": 7.0242656449553e-05, + "loss": 0.4373, + "step": 640 + }, + { + "epoch": 11.068965517241379, + "grad_norm": 13.804490089416504, + "learning_rate": 7.011494252873564e-05, + "loss": 0.7242, + "step": 642 + }, + { + "epoch": 11.10344827586207, + "grad_norm": 7.258790969848633, + "learning_rate": 6.998722860791826e-05, + "loss": 0.3412, + "step": 644 + }, + { + "epoch": 11.137931034482758, + "grad_norm": 13.48838996887207, + "learning_rate": 6.98595146871009e-05, + "loss": 0.7093, + "step": 646 + }, + { + "epoch": 11.172413793103448, + "grad_norm": 3.5530500411987305, + "learning_rate": 6.973180076628352e-05, + "loss": 0.1557, + "step": 648 + }, + { + "epoch": 11.206896551724139, + "grad_norm": 5.173003196716309, + "learning_rate": 6.960408684546615e-05, + "loss": 0.2043, + "step": 650 + }, + { + "epoch": 11.241379310344827, + "grad_norm": 8.948324203491211, + "learning_rate": 6.947637292464879e-05, + "loss": 0.3718, + "step": 652 + }, + { + "epoch": 11.275862068965518, + "grad_norm": 13.394603729248047, + "learning_rate": 6.934865900383142e-05, + "loss": 0.4858, + "step": 654 + }, + { + "epoch": 11.310344827586206, + "grad_norm": 3.575120210647583, + "learning_rate": 6.922094508301405e-05, + "loss": 0.2133, + "step": 656 + }, + { + "epoch": 11.344827586206897, + "grad_norm": 6.710623264312744, + "learning_rate": 6.909323116219668e-05, + "loss": 0.4072, + "step": 658 + }, + { + "epoch": 11.379310344827585, + "grad_norm": 14.04590892791748, + "learning_rate": 6.896551724137931e-05, + "loss": 0.9938, + "step": 660 + }, + { + "epoch": 11.413793103448276, + "grad_norm": 7.6388444900512695, + "learning_rate": 6.883780332056194e-05, + "loss": 0.4864, + "step": 662 + }, + { + "epoch": 11.448275862068966, + "grad_norm": 6.261629104614258, + "learning_rate": 6.871008939974458e-05, + "loss": 0.3112, + "step": 664 + }, + { + "epoch": 11.482758620689655, + "grad_norm": 7.990070819854736, + "learning_rate": 6.858237547892721e-05, + "loss": 0.6052, + "step": 666 + }, + { + "epoch": 11.517241379310345, + "grad_norm": 5.1240949630737305, + "learning_rate": 6.845466155810984e-05, + "loss": 0.6029, + "step": 668 + }, + { + "epoch": 11.551724137931034, + "grad_norm": 3.3157217502593994, + "learning_rate": 6.832694763729247e-05, + "loss": 0.2027, + "step": 670 + }, + { + "epoch": 11.586206896551724, + "grad_norm": 9.071480751037598, + "learning_rate": 6.81992337164751e-05, + "loss": 0.257, + "step": 672 + }, + { + "epoch": 11.620689655172415, + "grad_norm": 13.492743492126465, + "learning_rate": 6.807151979565773e-05, + "loss": 0.4914, + "step": 674 + }, + { + "epoch": 11.655172413793103, + "grad_norm": 1.5881291627883911, + "learning_rate": 6.794380587484037e-05, + "loss": 0.3057, + "step": 676 + }, + { + "epoch": 11.689655172413794, + "grad_norm": 13.12863540649414, + "learning_rate": 6.781609195402298e-05, + "loss": 0.4456, + "step": 678 + }, + { + "epoch": 11.724137931034482, + "grad_norm": 15.156206130981445, + "learning_rate": 6.768837803320563e-05, + "loss": 0.4095, + "step": 680 + }, + { + "epoch": 11.758620689655173, + "grad_norm": 12.477018356323242, + "learning_rate": 6.756066411238825e-05, + "loss": 0.6728, + "step": 682 + }, + { + "epoch": 11.793103448275861, + "grad_norm": 6.536929607391357, + "learning_rate": 6.74329501915709e-05, + "loss": 0.3572, + "step": 684 + }, + { + "epoch": 11.827586206896552, + "grad_norm": 8.295530319213867, + "learning_rate": 6.730523627075351e-05, + "loss": 0.3506, + "step": 686 + }, + { + "epoch": 11.862068965517242, + "grad_norm": 2.183727979660034, + "learning_rate": 6.717752234993616e-05, + "loss": 0.3145, + "step": 688 + }, + { + "epoch": 11.89655172413793, + "grad_norm": 9.358091354370117, + "learning_rate": 6.704980842911877e-05, + "loss": 0.5877, + "step": 690 + }, + { + "epoch": 11.931034482758621, + "grad_norm": 7.218227386474609, + "learning_rate": 6.69220945083014e-05, + "loss": 0.4683, + "step": 692 + }, + { + "epoch": 11.96551724137931, + "grad_norm": 6.784433364868164, + "learning_rate": 6.679438058748404e-05, + "loss": 0.3328, + "step": 694 + }, + { + "epoch": 12.0, + "grad_norm": 4.573869228363037, + "learning_rate": 6.666666666666667e-05, + "loss": 0.2319, + "step": 696 + }, + { + "epoch": 12.0, + "eval_accuracy": 0.6742424242424242, + "eval_f1_macro": 0.546672932330827, + "eval_f1_micro": 0.6742424242424242, + "eval_f1_weighted": 0.6409964779372674, + "eval_loss": 1.1553009748458862, + "eval_precision_macro": 0.5816405209018068, + "eval_precision_micro": 0.6742424242424242, + "eval_precision_weighted": 0.6699464486928733, + "eval_recall_macro": 0.5711466165413533, + "eval_recall_micro": 0.6742424242424242, + "eval_recall_weighted": 0.6742424242424242, + "eval_runtime": 1.9872, + "eval_samples_per_second": 66.426, + "eval_steps_per_second": 8.555, + "step": 696 + }, + { + "epoch": 12.03448275862069, + "grad_norm": 1.5767076015472412, + "learning_rate": 6.65389527458493e-05, + "loss": 0.235, + "step": 698 + }, + { + "epoch": 12.068965517241379, + "grad_norm": 7.686869144439697, + "learning_rate": 6.641123882503193e-05, + "loss": 0.5692, + "step": 700 + }, + { + "epoch": 12.10344827586207, + "grad_norm": 13.89122200012207, + "learning_rate": 6.628352490421456e-05, + "loss": 0.4877, + "step": 702 + }, + { + "epoch": 12.137931034482758, + "grad_norm": 4.811250686645508, + "learning_rate": 6.61558109833972e-05, + "loss": 0.2121, + "step": 704 + }, + { + "epoch": 12.172413793103448, + "grad_norm": 2.0311198234558105, + "learning_rate": 6.602809706257981e-05, + "loss": 0.2037, + "step": 706 + }, + { + "epoch": 12.206896551724139, + "grad_norm": 7.1766133308410645, + "learning_rate": 6.590038314176246e-05, + "loss": 0.2169, + "step": 708 + }, + { + "epoch": 12.241379310344827, + "grad_norm": 5.523502826690674, + "learning_rate": 6.577266922094508e-05, + "loss": 1.335, + "step": 710 + }, + { + "epoch": 12.275862068965518, + "grad_norm": 16.38228416442871, + "learning_rate": 6.564495530012772e-05, + "loss": 0.7502, + "step": 712 + }, + { + "epoch": 12.310344827586206, + "grad_norm": 8.219087600708008, + "learning_rate": 6.551724137931034e-05, + "loss": 0.7137, + "step": 714 + }, + { + "epoch": 12.344827586206897, + "grad_norm": 4.552286148071289, + "learning_rate": 6.538952745849299e-05, + "loss": 0.2233, + "step": 716 + }, + { + "epoch": 12.379310344827585, + "grad_norm": 2.2111048698425293, + "learning_rate": 6.52618135376756e-05, + "loss": 0.1856, + "step": 718 + }, + { + "epoch": 12.413793103448276, + "grad_norm": 10.02652359008789, + "learning_rate": 6.513409961685824e-05, + "loss": 0.4808, + "step": 720 + }, + { + "epoch": 12.448275862068966, + "grad_norm": 4.486725807189941, + "learning_rate": 6.500638569604087e-05, + "loss": 0.4057, + "step": 722 + }, + { + "epoch": 12.482758620689655, + "grad_norm": 1.8516465425491333, + "learning_rate": 6.48786717752235e-05, + "loss": 0.1974, + "step": 724 + }, + { + "epoch": 12.517241379310345, + "grad_norm": 6.200466156005859, + "learning_rate": 6.475095785440613e-05, + "loss": 0.2568, + "step": 726 + }, + { + "epoch": 12.551724137931034, + "grad_norm": 16.36850929260254, + "learning_rate": 6.462324393358876e-05, + "loss": 0.5874, + "step": 728 + }, + { + "epoch": 12.586206896551724, + "grad_norm": 4.7899346351623535, + "learning_rate": 6.44955300127714e-05, + "loss": 0.3983, + "step": 730 + }, + { + "epoch": 12.620689655172415, + "grad_norm": 7.357030868530273, + "learning_rate": 6.436781609195403e-05, + "loss": 0.3513, + "step": 732 + }, + { + "epoch": 12.655172413793103, + "grad_norm": 5.966126441955566, + "learning_rate": 6.424010217113666e-05, + "loss": 0.4016, + "step": 734 + }, + { + "epoch": 12.689655172413794, + "grad_norm": 6.382541656494141, + "learning_rate": 6.411238825031929e-05, + "loss": 0.1682, + "step": 736 + }, + { + "epoch": 12.724137931034482, + "grad_norm": 8.352551460266113, + "learning_rate": 6.398467432950191e-05, + "loss": 0.3286, + "step": 738 + }, + { + "epoch": 12.758620689655173, + "grad_norm": 9.189525604248047, + "learning_rate": 6.385696040868455e-05, + "loss": 0.49, + "step": 740 + }, + { + "epoch": 12.793103448275861, + "grad_norm": 15.152857780456543, + "learning_rate": 6.372924648786717e-05, + "loss": 0.1309, + "step": 742 + }, + { + "epoch": 12.827586206896552, + "grad_norm": 3.6843764781951904, + "learning_rate": 6.360153256704982e-05, + "loss": 0.1974, + "step": 744 + }, + { + "epoch": 12.862068965517242, + "grad_norm": 10.768837928771973, + "learning_rate": 6.347381864623243e-05, + "loss": 0.8215, + "step": 746 + }, + { + "epoch": 12.89655172413793, + "grad_norm": 7.068192005157471, + "learning_rate": 6.334610472541508e-05, + "loss": 0.2304, + "step": 748 + }, + { + "epoch": 12.931034482758621, + "grad_norm": 12.653234481811523, + "learning_rate": 6.32183908045977e-05, + "loss": 0.5768, + "step": 750 + }, + { + "epoch": 12.96551724137931, + "grad_norm": 7.09431266784668, + "learning_rate": 6.309067688378033e-05, + "loss": 0.6238, + "step": 752 + }, + { + "epoch": 13.0, + "grad_norm": 7.309642791748047, + "learning_rate": 6.296296296296296e-05, + "loss": 0.3528, + "step": 754 + }, + { + "epoch": 13.0, + "eval_accuracy": 0.6893939393939394, + "eval_f1_macro": 0.5794293474965745, + "eval_f1_micro": 0.6893939393939394, + "eval_f1_weighted": 0.6711440544996694, + "eval_loss": 1.1685433387756348, + "eval_precision_macro": 0.5887005339770317, + "eval_precision_micro": 0.6893939393939394, + "eval_precision_weighted": 0.6752094679514034, + "eval_recall_macro": 0.5955043859649124, + "eval_recall_micro": 0.6893939393939394, + "eval_recall_weighted": 0.6893939393939394, + "eval_runtime": 1.9392, + "eval_samples_per_second": 68.07, + "eval_steps_per_second": 8.767, + "step": 754 + }, + { + "epoch": 13.03448275862069, + "grad_norm": 9.388544082641602, + "learning_rate": 6.283524904214559e-05, + "loss": 0.6725, + "step": 756 + }, + { + "epoch": 13.068965517241379, + "grad_norm": 15.097204208374023, + "learning_rate": 6.270753512132822e-05, + "loss": 0.6723, + "step": 758 + }, + { + "epoch": 13.10344827586207, + "grad_norm": 7.6261796951293945, + "learning_rate": 6.257982120051086e-05, + "loss": 0.3793, + "step": 760 + }, + { + "epoch": 13.137931034482758, + "grad_norm": 2.728222370147705, + "learning_rate": 6.245210727969349e-05, + "loss": 0.1291, + "step": 762 + }, + { + "epoch": 13.172413793103448, + "grad_norm": 7.391420841217041, + "learning_rate": 6.232439335887612e-05, + "loss": 0.4385, + "step": 764 + }, + { + "epoch": 13.206896551724139, + "grad_norm": 9.982900619506836, + "learning_rate": 6.219667943805875e-05, + "loss": 0.3353, + "step": 766 + }, + { + "epoch": 13.241379310344827, + "grad_norm": 7.945486545562744, + "learning_rate": 6.206896551724138e-05, + "loss": 0.1623, + "step": 768 + }, + { + "epoch": 13.275862068965518, + "grad_norm": 11.18921184539795, + "learning_rate": 6.194125159642401e-05, + "loss": 0.1834, + "step": 770 + }, + { + "epoch": 13.310344827586206, + "grad_norm": 1.6263447999954224, + "learning_rate": 6.181353767560665e-05, + "loss": 0.0572, + "step": 772 + }, + { + "epoch": 13.344827586206897, + "grad_norm": 8.16580581665039, + "learning_rate": 6.168582375478928e-05, + "loss": 0.3301, + "step": 774 + }, + { + "epoch": 13.379310344827585, + "grad_norm": 2.9834048748016357, + "learning_rate": 6.155810983397191e-05, + "loss": 0.1367, + "step": 776 + }, + { + "epoch": 13.413793103448276, + "grad_norm": 9.801024436950684, + "learning_rate": 6.143039591315454e-05, + "loss": 0.2472, + "step": 778 + }, + { + "epoch": 13.448275862068966, + "grad_norm": 13.256599426269531, + "learning_rate": 6.130268199233716e-05, + "loss": 0.6242, + "step": 780 + }, + { + "epoch": 13.482758620689655, + "grad_norm": 12.600865364074707, + "learning_rate": 6.11749680715198e-05, + "loss": 0.387, + "step": 782 + }, + { + "epoch": 13.517241379310345, + "grad_norm": 12.394946098327637, + "learning_rate": 6.104725415070242e-05, + "loss": 0.3973, + "step": 784 + }, + { + "epoch": 13.551724137931034, + "grad_norm": 16.835023880004883, + "learning_rate": 6.091954022988506e-05, + "loss": 0.5736, + "step": 786 + }, + { + "epoch": 13.586206896551724, + "grad_norm": 10.348804473876953, + "learning_rate": 6.0791826309067686e-05, + "loss": 0.3203, + "step": 788 + }, + { + "epoch": 13.620689655172415, + "grad_norm": 10.827553749084473, + "learning_rate": 6.0664112388250325e-05, + "loss": 0.3311, + "step": 790 + }, + { + "epoch": 13.655172413793103, + "grad_norm": 4.213782787322998, + "learning_rate": 6.053639846743295e-05, + "loss": 0.0665, + "step": 792 + }, + { + "epoch": 13.689655172413794, + "grad_norm": 5.154265403747559, + "learning_rate": 6.040868454661558e-05, + "loss": 0.3031, + "step": 794 + }, + { + "epoch": 13.724137931034482, + "grad_norm": 18.196022033691406, + "learning_rate": 6.028097062579821e-05, + "loss": 0.8001, + "step": 796 + }, + { + "epoch": 13.758620689655173, + "grad_norm": 3.6230857372283936, + "learning_rate": 6.0153256704980845e-05, + "loss": 0.1406, + "step": 798 + }, + { + "epoch": 13.793103448275861, + "grad_norm": 0.9606171250343323, + "learning_rate": 6.0025542784163477e-05, + "loss": 0.0544, + "step": 800 + }, + { + "epoch": 13.827586206896552, + "grad_norm": 13.326411247253418, + "learning_rate": 5.989782886334611e-05, + "loss": 0.2442, + "step": 802 + }, + { + "epoch": 13.862068965517242, + "grad_norm": 8.518882751464844, + "learning_rate": 5.977011494252874e-05, + "loss": 0.2051, + "step": 804 + }, + { + "epoch": 13.89655172413793, + "grad_norm": 4.24729585647583, + "learning_rate": 5.964240102171137e-05, + "loss": 0.4868, + "step": 806 + }, + { + "epoch": 13.931034482758621, + "grad_norm": 0.7721981406211853, + "learning_rate": 5.9514687100893996e-05, + "loss": 0.1428, + "step": 808 + }, + { + "epoch": 13.96551724137931, + "grad_norm": 11.86274242401123, + "learning_rate": 5.9386973180076635e-05, + "loss": 0.7225, + "step": 810 + }, + { + "epoch": 14.0, + "grad_norm": 14.592938423156738, + "learning_rate": 5.925925925925926e-05, + "loss": 0.6238, + "step": 812 + }, + { + "epoch": 14.0, + "eval_accuracy": 0.6439393939393939, + "eval_f1_macro": 0.5579426867483074, + "eval_f1_micro": 0.6439393939393939, + "eval_f1_weighted": 0.6285409584589912, + "eval_loss": 1.1781333684921265, + "eval_precision_macro": 0.5450511546824911, + "eval_precision_micro": 0.6439393939393939, + "eval_precision_weighted": 0.6277759353712432, + "eval_recall_macro": 0.5856046365914788, + "eval_recall_micro": 0.6439393939393939, + "eval_recall_weighted": 0.6439393939393939, + "eval_runtime": 1.9417, + "eval_samples_per_second": 67.982, + "eval_steps_per_second": 8.755, + "step": 812 + }, + { + "epoch": 14.03448275862069, + "grad_norm": 11.65465259552002, + "learning_rate": 5.91315453384419e-05, + "loss": 0.3351, + "step": 814 + }, + { + "epoch": 14.068965517241379, + "grad_norm": 9.413543701171875, + "learning_rate": 5.900383141762452e-05, + "loss": 0.3187, + "step": 816 + }, + { + "epoch": 14.10344827586207, + "grad_norm": 3.177682638168335, + "learning_rate": 5.887611749680716e-05, + "loss": 0.0781, + "step": 818 + }, + { + "epoch": 14.137931034482758, + "grad_norm": 9.098755836486816, + "learning_rate": 5.8748403575989787e-05, + "loss": 0.1169, + "step": 820 + }, + { + "epoch": 14.172413793103448, + "grad_norm": 1.7022143602371216, + "learning_rate": 5.862068965517241e-05, + "loss": 0.2369, + "step": 822 + }, + { + "epoch": 14.206896551724139, + "grad_norm": 6.126805782318115, + "learning_rate": 5.849297573435505e-05, + "loss": 0.3991, + "step": 824 + }, + { + "epoch": 14.241379310344827, + "grad_norm": 6.332644939422607, + "learning_rate": 5.8365261813537675e-05, + "loss": 0.1479, + "step": 826 + }, + { + "epoch": 14.275862068965518, + "grad_norm": 5.615021228790283, + "learning_rate": 5.823754789272031e-05, + "loss": 0.1755, + "step": 828 + }, + { + "epoch": 14.310344827586206, + "grad_norm": 12.874805450439453, + "learning_rate": 5.810983397190294e-05, + "loss": 0.4218, + "step": 830 + }, + { + "epoch": 14.344827586206897, + "grad_norm": 0.3466058671474457, + "learning_rate": 5.798212005108558e-05, + "loss": 0.1017, + "step": 832 + }, + { + "epoch": 14.379310344827585, + "grad_norm": 0.1405513435602188, + "learning_rate": 5.78544061302682e-05, + "loss": 0.0431, + "step": 834 + }, + { + "epoch": 14.413793103448276, + "grad_norm": 2.837080717086792, + "learning_rate": 5.7726692209450826e-05, + "loss": 0.1162, + "step": 836 + }, + { + "epoch": 14.448275862068966, + "grad_norm": 0.49279162287712097, + "learning_rate": 5.7598978288633465e-05, + "loss": 0.094, + "step": 838 + }, + { + "epoch": 14.482758620689655, + "grad_norm": 3.950249671936035, + "learning_rate": 5.747126436781609e-05, + "loss": 0.2088, + "step": 840 + }, + { + "epoch": 14.517241379310345, + "grad_norm": 9.883584022521973, + "learning_rate": 5.734355044699873e-05, + "loss": 0.2122, + "step": 842 + }, + { + "epoch": 14.551724137931034, + "grad_norm": 8.051910400390625, + "learning_rate": 5.721583652618135e-05, + "loss": 0.3278, + "step": 844 + }, + { + "epoch": 14.586206896551724, + "grad_norm": 7.80661678314209, + "learning_rate": 5.708812260536399e-05, + "loss": 0.1348, + "step": 846 + }, + { + "epoch": 14.620689655172415, + "grad_norm": 6.488097190856934, + "learning_rate": 5.6960408684546617e-05, + "loss": 0.3984, + "step": 848 + }, + { + "epoch": 14.655172413793103, + "grad_norm": 12.74706745147705, + "learning_rate": 5.683269476372924e-05, + "loss": 0.3044, + "step": 850 + }, + { + "epoch": 14.689655172413794, + "grad_norm": 7.486535549163818, + "learning_rate": 5.670498084291188e-05, + "loss": 0.3548, + "step": 852 + }, + { + "epoch": 14.724137931034482, + "grad_norm": 4.710317611694336, + "learning_rate": 5.6577266922094505e-05, + "loss": 0.131, + "step": 854 + }, + { + "epoch": 14.758620689655173, + "grad_norm": 0.45310112833976746, + "learning_rate": 5.644955300127714e-05, + "loss": 0.2907, + "step": 856 + }, + { + "epoch": 14.793103448275861, + "grad_norm": 11.309956550598145, + "learning_rate": 5.632183908045977e-05, + "loss": 0.3132, + "step": 858 + }, + { + "epoch": 14.827586206896552, + "grad_norm": 16.222618103027344, + "learning_rate": 5.6194125159642407e-05, + "loss": 0.3499, + "step": 860 + }, + { + "epoch": 14.862068965517242, + "grad_norm": 19.321205139160156, + "learning_rate": 5.606641123882503e-05, + "loss": 1.193, + "step": 862 + }, + { + "epoch": 14.89655172413793, + "grad_norm": 14.328798294067383, + "learning_rate": 5.593869731800766e-05, + "loss": 0.1844, + "step": 864 + }, + { + "epoch": 14.931034482758621, + "grad_norm": 14.36063289642334, + "learning_rate": 5.5810983397190295e-05, + "loss": 1.2338, + "step": 866 + }, + { + "epoch": 14.96551724137931, + "grad_norm": 2.4482498168945312, + "learning_rate": 5.5683269476372927e-05, + "loss": 0.0565, + "step": 868 + }, + { + "epoch": 15.0, + "grad_norm": 9.643209457397461, + "learning_rate": 5.555555555555556e-05, + "loss": 0.1869, + "step": 870 + }, + { + "epoch": 15.0, + "eval_accuracy": 0.6060606060606061, + "eval_f1_macro": 0.5145970477437697, + "eval_f1_micro": 0.6060606060606061, + "eval_f1_weighted": 0.5982909919119969, + "eval_loss": 1.2305328845977783, + "eval_precision_macro": 0.5032325523840949, + "eval_precision_micro": 0.6060606060606061, + "eval_precision_weighted": 0.6012548090624109, + "eval_recall_macro": 0.5369360902255639, + "eval_recall_micro": 0.6060606060606061, + "eval_recall_weighted": 0.6060606060606061, + "eval_runtime": 1.9435, + "eval_samples_per_second": 67.917, + "eval_steps_per_second": 8.747, + "step": 870 + }, + { + "epoch": 15.03448275862069, + "grad_norm": 6.357883453369141, + "learning_rate": 5.542784163473819e-05, + "loss": 0.3321, + "step": 872 + }, + { + "epoch": 15.068965517241379, + "grad_norm": 13.48095989227295, + "learning_rate": 5.530012771392082e-05, + "loss": 0.1638, + "step": 874 + }, + { + "epoch": 15.10344827586207, + "grad_norm": 9.955169677734375, + "learning_rate": 5.517241379310345e-05, + "loss": 0.2777, + "step": 876 + }, + { + "epoch": 15.137931034482758, + "grad_norm": 6.751859664916992, + "learning_rate": 5.504469987228608e-05, + "loss": 0.1799, + "step": 878 + }, + { + "epoch": 15.172413793103448, + "grad_norm": 4.653921127319336, + "learning_rate": 5.491698595146872e-05, + "loss": 0.2553, + "step": 880 + }, + { + "epoch": 15.206896551724139, + "grad_norm": 5.335418224334717, + "learning_rate": 5.478927203065134e-05, + "loss": 0.1899, + "step": 882 + }, + { + "epoch": 15.241379310344827, + "grad_norm": 5.713738441467285, + "learning_rate": 5.466155810983398e-05, + "loss": 0.4833, + "step": 884 + }, + { + "epoch": 15.275862068965518, + "grad_norm": 5.765931129455566, + "learning_rate": 5.4533844189016605e-05, + "loss": 0.1149, + "step": 886 + }, + { + "epoch": 15.310344827586206, + "grad_norm": 8.66435718536377, + "learning_rate": 5.440613026819924e-05, + "loss": 0.3092, + "step": 888 + }, + { + "epoch": 15.344827586206897, + "grad_norm": 2.088322877883911, + "learning_rate": 5.427841634738187e-05, + "loss": 0.1967, + "step": 890 + }, + { + "epoch": 15.379310344827585, + "grad_norm": 12.177457809448242, + "learning_rate": 5.415070242656451e-05, + "loss": 0.1026, + "step": 892 + }, + { + "epoch": 15.413793103448276, + "grad_norm": 10.64006233215332, + "learning_rate": 5.402298850574713e-05, + "loss": 0.2939, + "step": 894 + }, + { + "epoch": 15.448275862068966, + "grad_norm": 18.743099212646484, + "learning_rate": 5.3895274584929756e-05, + "loss": 0.2609, + "step": 896 + }, + { + "epoch": 15.482758620689655, + "grad_norm": 10.08409309387207, + "learning_rate": 5.3767560664112395e-05, + "loss": 0.1403, + "step": 898 + }, + { + "epoch": 15.517241379310345, + "grad_norm": 10.607285499572754, + "learning_rate": 5.363984674329502e-05, + "loss": 0.2563, + "step": 900 + }, + { + "epoch": 15.551724137931034, + "grad_norm": 2.387784004211426, + "learning_rate": 5.351213282247766e-05, + "loss": 0.0304, + "step": 902 + }, + { + "epoch": 15.586206896551724, + "grad_norm": 1.1395808458328247, + "learning_rate": 5.338441890166028e-05, + "loss": 0.0853, + "step": 904 + }, + { + "epoch": 15.620689655172415, + "grad_norm": 7.147764682769775, + "learning_rate": 5.325670498084292e-05, + "loss": 0.3563, + "step": 906 + }, + { + "epoch": 15.655172413793103, + "grad_norm": 4.958852767944336, + "learning_rate": 5.3128991060025547e-05, + "loss": 0.1273, + "step": 908 + }, + { + "epoch": 15.689655172413794, + "grad_norm": 6.212413311004639, + "learning_rate": 5.300127713920817e-05, + "loss": 0.1748, + "step": 910 + }, + { + "epoch": 15.724137931034482, + "grad_norm": 1.9379197359085083, + "learning_rate": 5.287356321839081e-05, + "loss": 0.0297, + "step": 912 + }, + { + "epoch": 15.758620689655173, + "grad_norm": 1.3377583026885986, + "learning_rate": 5.2745849297573435e-05, + "loss": 0.0356, + "step": 914 + }, + { + "epoch": 15.793103448275861, + "grad_norm": 0.8058100938796997, + "learning_rate": 5.261813537675607e-05, + "loss": 0.1674, + "step": 916 + }, + { + "epoch": 15.827586206896552, + "grad_norm": 9.972665786743164, + "learning_rate": 5.24904214559387e-05, + "loss": 0.211, + "step": 918 + }, + { + "epoch": 15.862068965517242, + "grad_norm": 10.261574745178223, + "learning_rate": 5.236270753512134e-05, + "loss": 0.2202, + "step": 920 + }, + { + "epoch": 15.89655172413793, + "grad_norm": 4.782658576965332, + "learning_rate": 5.223499361430396e-05, + "loss": 0.1333, + "step": 922 + }, + { + "epoch": 15.931034482758621, + "grad_norm": 7.032027244567871, + "learning_rate": 5.2107279693486586e-05, + "loss": 0.2523, + "step": 924 + }, + { + "epoch": 15.96551724137931, + "grad_norm": 0.28720977902412415, + "learning_rate": 5.1979565772669225e-05, + "loss": 0.0323, + "step": 926 + }, + { + "epoch": 16.0, + "grad_norm": 9.86831283569336, + "learning_rate": 5.185185185185185e-05, + "loss": 0.1015, + "step": 928 + }, + { + "epoch": 16.0, + "eval_accuracy": 0.5909090909090909, + "eval_f1_macro": 0.501919425368257, + "eval_f1_micro": 0.5909090909090909, + "eval_f1_weighted": 0.5932466389209485, + "eval_loss": 1.3576174974441528, + "eval_precision_macro": 0.5439596385486938, + "eval_precision_micro": 0.5909090909090909, + "eval_precision_weighted": 0.6311543139507142, + "eval_recall_macro": 0.4959429824561404, + "eval_recall_micro": 0.5909090909090909, + "eval_recall_weighted": 0.5909090909090909, + "eval_runtime": 1.9838, + "eval_samples_per_second": 66.538, + "eval_steps_per_second": 8.569, + "step": 928 + }, + { + "epoch": 16.03448275862069, + "grad_norm": 2.869445323944092, + "learning_rate": 5.172413793103449e-05, + "loss": 0.3521, + "step": 930 + }, + { + "epoch": 16.06896551724138, + "grad_norm": 12.4180908203125, + "learning_rate": 5.159642401021711e-05, + "loss": 0.3301, + "step": 932 + }, + { + "epoch": 16.103448275862068, + "grad_norm": 3.6048808097839355, + "learning_rate": 5.146871008939975e-05, + "loss": 0.1339, + "step": 934 + }, + { + "epoch": 16.137931034482758, + "grad_norm": 0.6583413481712341, + "learning_rate": 5.1340996168582377e-05, + "loss": 0.3911, + "step": 936 + }, + { + "epoch": 16.17241379310345, + "grad_norm": 1.6497776508331299, + "learning_rate": 5.1213282247765e-05, + "loss": 0.2196, + "step": 938 + }, + { + "epoch": 16.20689655172414, + "grad_norm": 0.1662745475769043, + "learning_rate": 5.108556832694764e-05, + "loss": 0.0094, + "step": 940 + }, + { + "epoch": 16.24137931034483, + "grad_norm": 2.7962138652801514, + "learning_rate": 5.0957854406130265e-05, + "loss": 0.0605, + "step": 942 + }, + { + "epoch": 16.275862068965516, + "grad_norm": 5.833149433135986, + "learning_rate": 5.08301404853129e-05, + "loss": 0.2013, + "step": 944 + }, + { + "epoch": 16.310344827586206, + "grad_norm": 13.730070114135742, + "learning_rate": 5.070242656449553e-05, + "loss": 0.2639, + "step": 946 + }, + { + "epoch": 16.344827586206897, + "grad_norm": 5.876251220703125, + "learning_rate": 5.057471264367817e-05, + "loss": 0.3352, + "step": 948 + }, + { + "epoch": 16.379310344827587, + "grad_norm": 3.6209003925323486, + "learning_rate": 5.044699872286079e-05, + "loss": 0.1183, + "step": 950 + }, + { + "epoch": 16.413793103448278, + "grad_norm": 0.6604540944099426, + "learning_rate": 5.031928480204342e-05, + "loss": 0.0778, + "step": 952 + }, + { + "epoch": 16.448275862068964, + "grad_norm": 1.6417039632797241, + "learning_rate": 5.0191570881226055e-05, + "loss": 0.0523, + "step": 954 + }, + { + "epoch": 16.482758620689655, + "grad_norm": 4.871641159057617, + "learning_rate": 5.0063856960408687e-05, + "loss": 0.1488, + "step": 956 + }, + { + "epoch": 16.517241379310345, + "grad_norm": 0.6309295892715454, + "learning_rate": 4.993614303959132e-05, + "loss": 0.039, + "step": 958 + }, + { + "epoch": 16.551724137931036, + "grad_norm": 9.119877815246582, + "learning_rate": 4.980842911877395e-05, + "loss": 0.1025, + "step": 960 + }, + { + "epoch": 16.586206896551722, + "grad_norm": 1.3798292875289917, + "learning_rate": 4.968071519795658e-05, + "loss": 0.0307, + "step": 962 + }, + { + "epoch": 16.620689655172413, + "grad_norm": 0.3394834101200104, + "learning_rate": 4.955300127713921e-05, + "loss": 0.04, + "step": 964 + }, + { + "epoch": 16.655172413793103, + "grad_norm": 0.6206883788108826, + "learning_rate": 4.9425287356321845e-05, + "loss": 0.1483, + "step": 966 + }, + { + "epoch": 16.689655172413794, + "grad_norm": 9.715845108032227, + "learning_rate": 4.929757343550448e-05, + "loss": 0.3286, + "step": 968 + }, + { + "epoch": 16.724137931034484, + "grad_norm": 1.0101009607315063, + "learning_rate": 4.916985951468711e-05, + "loss": 0.158, + "step": 970 + }, + { + "epoch": 16.75862068965517, + "grad_norm": 17.89195442199707, + "learning_rate": 4.904214559386973e-05, + "loss": 0.0978, + "step": 972 + }, + { + "epoch": 16.79310344827586, + "grad_norm": 3.084064245223999, + "learning_rate": 4.8914431673052365e-05, + "loss": 0.0461, + "step": 974 + }, + { + "epoch": 16.82758620689655, + "grad_norm": 0.12563039362430573, + "learning_rate": 4.8786717752234997e-05, + "loss": 0.0234, + "step": 976 + }, + { + "epoch": 16.862068965517242, + "grad_norm": 1.2776066064834595, + "learning_rate": 4.865900383141763e-05, + "loss": 0.0931, + "step": 978 + }, + { + "epoch": 16.896551724137932, + "grad_norm": 2.729607343673706, + "learning_rate": 4.853128991060026e-05, + "loss": 0.0507, + "step": 980 + }, + { + "epoch": 16.93103448275862, + "grad_norm": 0.35684946179389954, + "learning_rate": 4.840357598978289e-05, + "loss": 0.1429, + "step": 982 + }, + { + "epoch": 16.96551724137931, + "grad_norm": 4.096828460693359, + "learning_rate": 4.827586206896552e-05, + "loss": 0.321, + "step": 984 + }, + { + "epoch": 17.0, + "grad_norm": 0.25383704900741577, + "learning_rate": 4.814814814814815e-05, + "loss": 0.3809, + "step": 986 + }, + { + "epoch": 17.0, + "eval_accuracy": 0.6590909090909091, + "eval_f1_macro": 0.5666514454561888, + "eval_f1_micro": 0.6590909090909091, + "eval_f1_weighted": 0.652650863169293, + "eval_loss": 1.2998155355453491, + "eval_precision_macro": 0.5827908086311447, + "eval_precision_micro": 0.6590909090909091, + "eval_precision_weighted": 0.6885351168439403, + "eval_recall_macro": 0.5838189223057644, + "eval_recall_micro": 0.6590909090909091, + "eval_recall_weighted": 0.6590909090909091, + "eval_runtime": 2.7743, + "eval_samples_per_second": 47.58, + "eval_steps_per_second": 6.128, + "step": 986 + }, + { + "epoch": 17.03448275862069, + "grad_norm": 2.5796058177948, + "learning_rate": 4.802043422733078e-05, + "loss": 0.0584, + "step": 988 + }, + { + "epoch": 17.06896551724138, + "grad_norm": 0.15105970203876495, + "learning_rate": 4.789272030651341e-05, + "loss": 0.0102, + "step": 990 + }, + { + "epoch": 17.103448275862068, + "grad_norm": 1.5413386821746826, + "learning_rate": 4.776500638569604e-05, + "loss": 0.039, + "step": 992 + }, + { + "epoch": 17.137931034482758, + "grad_norm": 20.432376861572266, + "learning_rate": 4.7637292464878675e-05, + "loss": 0.1994, + "step": 994 + }, + { + "epoch": 17.17241379310345, + "grad_norm": 13.159724235534668, + "learning_rate": 4.7509578544061307e-05, + "loss": 0.1881, + "step": 996 + }, + { + "epoch": 17.20689655172414, + "grad_norm": 0.3454015552997589, + "learning_rate": 4.738186462324394e-05, + "loss": 0.0048, + "step": 998 + }, + { + "epoch": 17.24137931034483, + "grad_norm": 20.270763397216797, + "learning_rate": 4.725415070242657e-05, + "loss": 0.123, + "step": 1000 + }, + { + "epoch": 17.275862068965516, + "grad_norm": 22.238080978393555, + "learning_rate": 4.7126436781609195e-05, + "loss": 0.2302, + "step": 1002 + }, + { + "epoch": 17.310344827586206, + "grad_norm": 2.1673974990844727, + "learning_rate": 4.6998722860791827e-05, + "loss": 0.0187, + "step": 1004 + }, + { + "epoch": 17.344827586206897, + "grad_norm": 0.3968747556209564, + "learning_rate": 4.687100893997446e-05, + "loss": 0.0076, + "step": 1006 + }, + { + "epoch": 17.379310344827587, + "grad_norm": 0.5099461078643799, + "learning_rate": 4.674329501915709e-05, + "loss": 0.0134, + "step": 1008 + }, + { + "epoch": 17.413793103448278, + "grad_norm": 8.667265892028809, + "learning_rate": 4.661558109833972e-05, + "loss": 0.2712, + "step": 1010 + }, + { + "epoch": 17.448275862068964, + "grad_norm": 6.667606353759766, + "learning_rate": 4.648786717752235e-05, + "loss": 0.1044, + "step": 1012 + }, + { + "epoch": 17.482758620689655, + "grad_norm": 13.070887565612793, + "learning_rate": 4.6360153256704985e-05, + "loss": 0.3492, + "step": 1014 + }, + { + "epoch": 17.517241379310345, + "grad_norm": 4.142744064331055, + "learning_rate": 4.623243933588761e-05, + "loss": 0.0549, + "step": 1016 + }, + { + "epoch": 17.551724137931036, + "grad_norm": 0.10112213343381882, + "learning_rate": 4.610472541507024e-05, + "loss": 0.0567, + "step": 1018 + }, + { + "epoch": 17.586206896551722, + "grad_norm": 0.03555409982800484, + "learning_rate": 4.597701149425287e-05, + "loss": 0.0382, + "step": 1020 + }, + { + "epoch": 17.620689655172413, + "grad_norm": 0.5085342526435852, + "learning_rate": 4.5849297573435505e-05, + "loss": 0.0261, + "step": 1022 + }, + { + "epoch": 17.655172413793103, + "grad_norm": 1.362773060798645, + "learning_rate": 4.5721583652618137e-05, + "loss": 0.1, + "step": 1024 + }, + { + "epoch": 17.689655172413794, + "grad_norm": 2.903794050216675, + "learning_rate": 4.559386973180077e-05, + "loss": 0.0157, + "step": 1026 + }, + { + "epoch": 17.724137931034484, + "grad_norm": 0.29353153705596924, + "learning_rate": 4.54661558109834e-05, + "loss": 0.0185, + "step": 1028 + }, + { + "epoch": 17.75862068965517, + "grad_norm": 4.77042293548584, + "learning_rate": 4.5338441890166025e-05, + "loss": 0.0405, + "step": 1030 + }, + { + "epoch": 17.79310344827586, + "grad_norm": 2.864474296569824, + "learning_rate": 4.5210727969348656e-05, + "loss": 0.3131, + "step": 1032 + }, + { + "epoch": 17.82758620689655, + "grad_norm": 1.4102516174316406, + "learning_rate": 4.508301404853129e-05, + "loss": 0.1175, + "step": 1034 + }, + { + "epoch": 17.862068965517242, + "grad_norm": 5.741951942443848, + "learning_rate": 4.495530012771392e-05, + "loss": 0.3193, + "step": 1036 + }, + { + "epoch": 17.896551724137932, + "grad_norm": 0.3935476243495941, + "learning_rate": 4.482758620689655e-05, + "loss": 0.0074, + "step": 1038 + }, + { + "epoch": 17.93103448275862, + "grad_norm": 6.773815155029297, + "learning_rate": 4.469987228607918e-05, + "loss": 0.0759, + "step": 1040 + }, + { + "epoch": 17.96551724137931, + "grad_norm": 14.86123275756836, + "learning_rate": 4.4572158365261815e-05, + "loss": 0.1914, + "step": 1042 + }, + { + "epoch": 18.0, + "grad_norm": 5.217438697814941, + "learning_rate": 4.4444444444444447e-05, + "loss": 0.0887, + "step": 1044 + }, + { + "epoch": 18.0, + "eval_accuracy": 0.6666666666666666, + "eval_f1_macro": 0.5571781194602418, + "eval_f1_micro": 0.6666666666666666, + "eval_f1_weighted": 0.6489059749380235, + "eval_loss": 1.4153717756271362, + "eval_precision_macro": 0.5682278640899331, + "eval_precision_micro": 0.6666666666666666, + "eval_precision_weighted": 0.6518116175012726, + "eval_recall_macro": 0.568342731829574, + "eval_recall_micro": 0.6666666666666666, + "eval_recall_weighted": 0.6666666666666666, + "eval_runtime": 2.0979, + "eval_samples_per_second": 62.919, + "eval_steps_per_second": 8.103, + "step": 1044 + }, + { + "epoch": 18.03448275862069, + "grad_norm": 15.641746520996094, + "learning_rate": 4.431673052362708e-05, + "loss": 0.1409, + "step": 1046 + }, + { + "epoch": 18.06896551724138, + "grad_norm": 11.60824203491211, + "learning_rate": 4.418901660280971e-05, + "loss": 0.4299, + "step": 1048 + }, + { + "epoch": 18.103448275862068, + "grad_norm": 0.08567314594984055, + "learning_rate": 4.406130268199234e-05, + "loss": 0.0062, + "step": 1050 + }, + { + "epoch": 18.137931034482758, + "grad_norm": 0.12726525962352753, + "learning_rate": 4.393358876117497e-05, + "loss": 0.0031, + "step": 1052 + }, + { + "epoch": 18.17241379310345, + "grad_norm": 0.3815951645374298, + "learning_rate": 4.3805874840357605e-05, + "loss": 0.004, + "step": 1054 + }, + { + "epoch": 18.20689655172414, + "grad_norm": 17.277345657348633, + "learning_rate": 4.367816091954024e-05, + "loss": 0.1928, + "step": 1056 + }, + { + "epoch": 18.24137931034483, + "grad_norm": 0.14995414018630981, + "learning_rate": 4.355044699872286e-05, + "loss": 0.1884, + "step": 1058 + }, + { + "epoch": 18.275862068965516, + "grad_norm": 0.1205284371972084, + "learning_rate": 4.342273307790549e-05, + "loss": 0.1822, + "step": 1060 + }, + { + "epoch": 18.310344827586206, + "grad_norm": 0.35154199600219727, + "learning_rate": 4.3295019157088125e-05, + "loss": 0.019, + "step": 1062 + }, + { + "epoch": 18.344827586206897, + "grad_norm": 4.302131652832031, + "learning_rate": 4.3167305236270757e-05, + "loss": 0.0265, + "step": 1064 + }, + { + "epoch": 18.379310344827587, + "grad_norm": 1.0111949443817139, + "learning_rate": 4.303959131545339e-05, + "loss": 0.0126, + "step": 1066 + }, + { + "epoch": 18.413793103448278, + "grad_norm": 12.860360145568848, + "learning_rate": 4.291187739463602e-05, + "loss": 0.1918, + "step": 1068 + }, + { + "epoch": 18.448275862068964, + "grad_norm": 1.465510368347168, + "learning_rate": 4.278416347381865e-05, + "loss": 0.1672, + "step": 1070 + }, + { + "epoch": 18.482758620689655, + "grad_norm": 0.1530478149652481, + "learning_rate": 4.2656449553001277e-05, + "loss": 0.0624, + "step": 1072 + }, + { + "epoch": 18.517241379310345, + "grad_norm": 7.125781059265137, + "learning_rate": 4.252873563218391e-05, + "loss": 0.359, + "step": 1074 + }, + { + "epoch": 18.551724137931036, + "grad_norm": 3.524441957473755, + "learning_rate": 4.240102171136654e-05, + "loss": 0.0263, + "step": 1076 + }, + { + "epoch": 18.586206896551722, + "grad_norm": 2.0102035999298096, + "learning_rate": 4.227330779054917e-05, + "loss": 0.0602, + "step": 1078 + }, + { + "epoch": 18.620689655172413, + "grad_norm": 0.3677597939968109, + "learning_rate": 4.21455938697318e-05, + "loss": 0.0251, + "step": 1080 + }, + { + "epoch": 18.655172413793103, + "grad_norm": 19.12733268737793, + "learning_rate": 4.2017879948914435e-05, + "loss": 0.4779, + "step": 1082 + }, + { + "epoch": 18.689655172413794, + "grad_norm": 2.7869174480438232, + "learning_rate": 4.189016602809707e-05, + "loss": 0.1591, + "step": 1084 + }, + { + "epoch": 18.724137931034484, + "grad_norm": 2.577024459838867, + "learning_rate": 4.17624521072797e-05, + "loss": 0.0862, + "step": 1086 + }, + { + "epoch": 18.75862068965517, + "grad_norm": 0.28967535495758057, + "learning_rate": 4.163473818646232e-05, + "loss": 0.0645, + "step": 1088 + }, + { + "epoch": 18.79310344827586, + "grad_norm": 15.79033088684082, + "learning_rate": 4.1507024265644955e-05, + "loss": 0.1082, + "step": 1090 + }, + { + "epoch": 18.82758620689655, + "grad_norm": 0.5589704513549805, + "learning_rate": 4.1379310344827587e-05, + "loss": 0.0205, + "step": 1092 + }, + { + "epoch": 18.862068965517242, + "grad_norm": 0.6776816844940186, + "learning_rate": 4.125159642401022e-05, + "loss": 0.0067, + "step": 1094 + }, + { + "epoch": 18.896551724137932, + "grad_norm": 0.27798280119895935, + "learning_rate": 4.112388250319285e-05, + "loss": 0.0039, + "step": 1096 + }, + { + "epoch": 18.93103448275862, + "grad_norm": 12.297942161560059, + "learning_rate": 4.099616858237548e-05, + "loss": 0.151, + "step": 1098 + }, + { + "epoch": 18.96551724137931, + "grad_norm": 2.289623737335205, + "learning_rate": 4.086845466155811e-05, + "loss": 0.176, + "step": 1100 + }, + { + "epoch": 19.0, + "grad_norm": 11.227123260498047, + "learning_rate": 4.074074074074074e-05, + "loss": 0.1422, + "step": 1102 + }, + { + "epoch": 19.0, + "eval_accuracy": 0.6666666666666666, + "eval_f1_macro": 0.5608736126029359, + "eval_f1_micro": 0.6666666666666666, + "eval_f1_weighted": 0.647220323536113, + "eval_loss": 1.398878574371338, + "eval_precision_macro": 0.5671867606078133, + "eval_precision_micro": 0.6666666666666666, + "eval_precision_weighted": 0.6419607605313348, + "eval_recall_macro": 0.5695332080200501, + "eval_recall_micro": 0.6666666666666666, + "eval_recall_weighted": 0.6666666666666666, + "eval_runtime": 3.3441, + "eval_samples_per_second": 39.472, + "eval_steps_per_second": 5.084, + "step": 1102 + }, + { + "epoch": 19.03448275862069, + "grad_norm": 3.0859837532043457, + "learning_rate": 4.061302681992337e-05, + "loss": 0.0213, + "step": 1104 + }, + { + "epoch": 19.06896551724138, + "grad_norm": 12.9458589553833, + "learning_rate": 4.0485312899106e-05, + "loss": 0.1051, + "step": 1106 + }, + { + "epoch": 19.103448275862068, + "grad_norm": 0.20263761281967163, + "learning_rate": 4.035759897828863e-05, + "loss": 0.0172, + "step": 1108 + }, + { + "epoch": 19.137931034482758, + "grad_norm": 0.19154572486877441, + "learning_rate": 4.0229885057471265e-05, + "loss": 0.035, + "step": 1110 + }, + { + "epoch": 19.17241379310345, + "grad_norm": 0.30079570412635803, + "learning_rate": 4.0102171136653897e-05, + "loss": 0.0363, + "step": 1112 + }, + { + "epoch": 19.20689655172414, + "grad_norm": 0.17321471869945526, + "learning_rate": 3.997445721583653e-05, + "loss": 0.0143, + "step": 1114 + }, + { + "epoch": 19.24137931034483, + "grad_norm": 0.6283549666404724, + "learning_rate": 3.984674329501916e-05, + "loss": 0.0288, + "step": 1116 + }, + { + "epoch": 19.275862068965516, + "grad_norm": 0.014564587734639645, + "learning_rate": 3.971902937420179e-05, + "loss": 0.005, + "step": 1118 + }, + { + "epoch": 19.310344827586206, + "grad_norm": 0.0865383893251419, + "learning_rate": 3.959131545338442e-05, + "loss": 0.0134, + "step": 1120 + }, + { + "epoch": 19.344827586206897, + "grad_norm": 0.13445672392845154, + "learning_rate": 3.9463601532567055e-05, + "loss": 0.0104, + "step": 1122 + }, + { + "epoch": 19.379310344827587, + "grad_norm": 1.2037104368209839, + "learning_rate": 3.933588761174969e-05, + "loss": 0.0126, + "step": 1124 + }, + { + "epoch": 19.413793103448278, + "grad_norm": 0.8391666412353516, + "learning_rate": 3.920817369093232e-05, + "loss": 0.0176, + "step": 1126 + }, + { + "epoch": 19.448275862068964, + "grad_norm": 11.825739860534668, + "learning_rate": 3.908045977011495e-05, + "loss": 0.0852, + "step": 1128 + }, + { + "epoch": 19.482758620689655, + "grad_norm": 4.42478609085083, + "learning_rate": 3.8952745849297575e-05, + "loss": 0.077, + "step": 1130 + }, + { + "epoch": 19.517241379310345, + "grad_norm": 0.5317286849021912, + "learning_rate": 3.8825031928480207e-05, + "loss": 0.0204, + "step": 1132 + }, + { + "epoch": 19.551724137931036, + "grad_norm": 0.39885902404785156, + "learning_rate": 3.869731800766284e-05, + "loss": 0.0064, + "step": 1134 + }, + { + "epoch": 19.586206896551722, + "grad_norm": 0.8824494481086731, + "learning_rate": 3.856960408684547e-05, + "loss": 0.0184, + "step": 1136 + }, + { + "epoch": 19.620689655172413, + "grad_norm": 1.4008420705795288, + "learning_rate": 3.84418901660281e-05, + "loss": 0.0778, + "step": 1138 + }, + { + "epoch": 19.655172413793103, + "grad_norm": 0.5302988290786743, + "learning_rate": 3.831417624521073e-05, + "loss": 0.0935, + "step": 1140 + }, + { + "epoch": 19.689655172413794, + "grad_norm": 12.098783493041992, + "learning_rate": 3.8186462324393365e-05, + "loss": 0.4228, + "step": 1142 + }, + { + "epoch": 19.724137931034484, + "grad_norm": 2.2826879024505615, + "learning_rate": 3.805874840357599e-05, + "loss": 0.0317, + "step": 1144 + }, + { + "epoch": 19.75862068965517, + "grad_norm": 0.4077673852443695, + "learning_rate": 3.793103448275862e-05, + "loss": 0.0078, + "step": 1146 + }, + { + "epoch": 19.79310344827586, + "grad_norm": 6.110368728637695, + "learning_rate": 3.780332056194125e-05, + "loss": 0.1554, + "step": 1148 + }, + { + "epoch": 19.82758620689655, + "grad_norm": 0.027513781562447548, + "learning_rate": 3.7675606641123885e-05, + "loss": 0.0625, + "step": 1150 + }, + { + "epoch": 19.862068965517242, + "grad_norm": 6.546108245849609, + "learning_rate": 3.7547892720306517e-05, + "loss": 0.0988, + "step": 1152 + }, + { + "epoch": 19.896551724137932, + "grad_norm": 0.08968118578195572, + "learning_rate": 3.742017879948915e-05, + "loss": 0.002, + "step": 1154 + }, + { + "epoch": 19.93103448275862, + "grad_norm": 0.14957240223884583, + "learning_rate": 3.729246487867178e-05, + "loss": 0.0024, + "step": 1156 + }, + { + "epoch": 19.96551724137931, + "grad_norm": 21.937902450561523, + "learning_rate": 3.716475095785441e-05, + "loss": 0.4017, + "step": 1158 + }, + { + "epoch": 20.0, + "grad_norm": 0.162915900349617, + "learning_rate": 3.7037037037037037e-05, + "loss": 0.0037, + "step": 1160 + }, + { + "epoch": 20.0, + "eval_accuracy": 0.6212121212121212, + "eval_f1_macro": 0.5241622523135129, + "eval_f1_micro": 0.6212121212121212, + "eval_f1_weighted": 0.6077529051459533, + "eval_loss": 1.5133600234985352, + "eval_precision_macro": 0.5262823725981621, + "eval_precision_micro": 0.6212121212121212, + "eval_precision_weighted": 0.6092636895268474, + "eval_recall_macro": 0.537390350877193, + "eval_recall_micro": 0.6212121212121212, + "eval_recall_weighted": 0.6212121212121212, + "eval_runtime": 2.2507, + "eval_samples_per_second": 58.65, + "eval_steps_per_second": 7.553, + "step": 1160 + }, + { + "epoch": 20.03448275862069, + "grad_norm": 0.17806491255760193, + "learning_rate": 3.690932311621967e-05, + "loss": 0.0142, + "step": 1162 + }, + { + "epoch": 20.06896551724138, + "grad_norm": 0.30130231380462646, + "learning_rate": 3.67816091954023e-05, + "loss": 0.0266, + "step": 1164 + }, + { + "epoch": 20.103448275862068, + "grad_norm": 0.30467385053634644, + "learning_rate": 3.665389527458493e-05, + "loss": 0.0289, + "step": 1166 + }, + { + "epoch": 20.137931034482758, + "grad_norm": 6.654253959655762, + "learning_rate": 3.652618135376756e-05, + "loss": 0.1494, + "step": 1168 + }, + { + "epoch": 20.17241379310345, + "grad_norm": 29.355484008789062, + "learning_rate": 3.6398467432950195e-05, + "loss": 0.3555, + "step": 1170 + }, + { + "epoch": 20.20689655172414, + "grad_norm": 0.25190678238868713, + "learning_rate": 3.627075351213283e-05, + "loss": 0.0065, + "step": 1172 + }, + { + "epoch": 20.24137931034483, + "grad_norm": 7.903636455535889, + "learning_rate": 3.614303959131545e-05, + "loss": 0.0348, + "step": 1174 + }, + { + "epoch": 20.275862068965516, + "grad_norm": 1.2426676750183105, + "learning_rate": 3.601532567049808e-05, + "loss": 0.0348, + "step": 1176 + }, + { + "epoch": 20.310344827586206, + "grad_norm": 0.05831276252865791, + "learning_rate": 3.5887611749680715e-05, + "loss": 0.0171, + "step": 1178 + }, + { + "epoch": 20.344827586206897, + "grad_norm": 3.8727033138275146, + "learning_rate": 3.5759897828863347e-05, + "loss": 0.2391, + "step": 1180 + }, + { + "epoch": 20.379310344827587, + "grad_norm": 2.7642955780029297, + "learning_rate": 3.563218390804598e-05, + "loss": 0.0218, + "step": 1182 + }, + { + "epoch": 20.413793103448278, + "grad_norm": 28.808361053466797, + "learning_rate": 3.550446998722861e-05, + "loss": 0.1445, + "step": 1184 + }, + { + "epoch": 20.448275862068964, + "grad_norm": 0.25608864426612854, + "learning_rate": 3.537675606641124e-05, + "loss": 0.0038, + "step": 1186 + }, + { + "epoch": 20.482758620689655, + "grad_norm": 0.18029409646987915, + "learning_rate": 3.5249042145593867e-05, + "loss": 0.0034, + "step": 1188 + }, + { + "epoch": 20.517241379310345, + "grad_norm": 0.03822282329201698, + "learning_rate": 3.51213282247765e-05, + "loss": 0.0623, + "step": 1190 + }, + { + "epoch": 20.551724137931036, + "grad_norm": 0.24421456456184387, + "learning_rate": 3.499361430395913e-05, + "loss": 0.0054, + "step": 1192 + }, + { + "epoch": 20.586206896551722, + "grad_norm": 13.820355415344238, + "learning_rate": 3.486590038314176e-05, + "loss": 0.0834, + "step": 1194 + }, + { + "epoch": 20.620689655172413, + "grad_norm": 3.458310127258301, + "learning_rate": 3.473818646232439e-05, + "loss": 0.0236, + "step": 1196 + }, + { + "epoch": 20.655172413793103, + "grad_norm": 0.02704853191971779, + "learning_rate": 3.4610472541507025e-05, + "loss": 0.0033, + "step": 1198 + }, + { + "epoch": 20.689655172413794, + "grad_norm": 1.1011626720428467, + "learning_rate": 3.4482758620689657e-05, + "loss": 0.0165, + "step": 1200 + }, + { + "epoch": 20.724137931034484, + "grad_norm": 21.770254135131836, + "learning_rate": 3.435504469987229e-05, + "loss": 0.1164, + "step": 1202 + }, + { + "epoch": 20.75862068965517, + "grad_norm": 0.3167782425880432, + "learning_rate": 3.422733077905492e-05, + "loss": 0.0036, + "step": 1204 + }, + { + "epoch": 20.79310344827586, + "grad_norm": 0.057486578822135925, + "learning_rate": 3.409961685823755e-05, + "loss": 0.0029, + "step": 1206 + }, + { + "epoch": 20.82758620689655, + "grad_norm": 0.7752865552902222, + "learning_rate": 3.397190293742018e-05, + "loss": 0.1632, + "step": 1208 + }, + { + "epoch": 20.862068965517242, + "grad_norm": 1.2659624814987183, + "learning_rate": 3.3844189016602815e-05, + "loss": 0.0084, + "step": 1210 + }, + { + "epoch": 20.896551724137932, + "grad_norm": 0.0808514654636383, + "learning_rate": 3.371647509578545e-05, + "loss": 0.0011, + "step": 1212 + }, + { + "epoch": 20.93103448275862, + "grad_norm": 0.502754807472229, + "learning_rate": 3.358876117496808e-05, + "loss": 0.0062, + "step": 1214 + }, + { + "epoch": 20.96551724137931, + "grad_norm": 0.09211990982294083, + "learning_rate": 3.34610472541507e-05, + "loss": 0.0043, + "step": 1216 + }, + { + "epoch": 21.0, + "grad_norm": 0.22179439663887024, + "learning_rate": 3.3333333333333335e-05, + "loss": 0.0602, + "step": 1218 + }, + { + "epoch": 21.0, + "eval_accuracy": 0.6666666666666666, + "eval_f1_macro": 0.5660444695559945, + "eval_f1_micro": 0.6666666666666666, + "eval_f1_weighted": 0.6544395105724697, + "eval_loss": 1.5348953008651733, + "eval_precision_macro": 0.5710371698009604, + "eval_precision_micro": 0.6666666666666666, + "eval_precision_weighted": 0.6503411219319607, + "eval_recall_macro": 0.5670739348370927, + "eval_recall_micro": 0.6666666666666666, + "eval_recall_weighted": 0.6666666666666666, + "eval_runtime": 2.9914, + "eval_samples_per_second": 44.127, + "eval_steps_per_second": 5.683, + "step": 1218 + }, + { + "epoch": 21.03448275862069, + "grad_norm": 0.02936175838112831, + "learning_rate": 3.3205619412515967e-05, + "loss": 0.0007, + "step": 1220 + }, + { + "epoch": 21.06896551724138, + "grad_norm": 10.468208312988281, + "learning_rate": 3.30779054916986e-05, + "loss": 0.0762, + "step": 1222 + }, + { + "epoch": 21.103448275862068, + "grad_norm": 0.010613945312798023, + "learning_rate": 3.295019157088123e-05, + "loss": 0.0008, + "step": 1224 + }, + { + "epoch": 21.137931034482758, + "grad_norm": 0.8747249841690063, + "learning_rate": 3.282247765006386e-05, + "loss": 0.3968, + "step": 1226 + }, + { + "epoch": 21.17241379310345, + "grad_norm": 0.014937883242964745, + "learning_rate": 3.269476372924649e-05, + "loss": 0.0526, + "step": 1228 + }, + { + "epoch": 21.20689655172414, + "grad_norm": 0.29066944122314453, + "learning_rate": 3.256704980842912e-05, + "loss": 0.0051, + "step": 1230 + }, + { + "epoch": 21.24137931034483, + "grad_norm": 0.05553491786122322, + "learning_rate": 3.243933588761175e-05, + "loss": 0.0015, + "step": 1232 + }, + { + "epoch": 21.275862068965516, + "grad_norm": 0.1087840124964714, + "learning_rate": 3.231162196679438e-05, + "loss": 0.0035, + "step": 1234 + }, + { + "epoch": 21.310344827586206, + "grad_norm": 2.526913642883301, + "learning_rate": 3.218390804597701e-05, + "loss": 0.0228, + "step": 1236 + }, + { + "epoch": 21.344827586206897, + "grad_norm": 0.09952912479639053, + "learning_rate": 3.2056194125159645e-05, + "loss": 0.0025, + "step": 1238 + }, + { + "epoch": 21.379310344827587, + "grad_norm": 0.03265494853258133, + "learning_rate": 3.192848020434228e-05, + "loss": 0.0044, + "step": 1240 + }, + { + "epoch": 21.413793103448278, + "grad_norm": 5.046082973480225, + "learning_rate": 3.180076628352491e-05, + "loss": 0.058, + "step": 1242 + }, + { + "epoch": 21.448275862068964, + "grad_norm": 15.606002807617188, + "learning_rate": 3.167305236270754e-05, + "loss": 0.1473, + "step": 1244 + }, + { + "epoch": 21.482758620689655, + "grad_norm": 19.64705467224121, + "learning_rate": 3.1545338441890165e-05, + "loss": 0.1913, + "step": 1246 + }, + { + "epoch": 21.517241379310345, + "grad_norm": 4.021636009216309, + "learning_rate": 3.1417624521072797e-05, + "loss": 0.0729, + "step": 1248 + }, + { + "epoch": 21.551724137931036, + "grad_norm": 0.22693483531475067, + "learning_rate": 3.128991060025543e-05, + "loss": 0.0402, + "step": 1250 + }, + { + "epoch": 21.586206896551722, + "grad_norm": 0.1609225869178772, + "learning_rate": 3.116219667943806e-05, + "loss": 0.0025, + "step": 1252 + }, + { + "epoch": 21.620689655172413, + "grad_norm": 0.10900776833295822, + "learning_rate": 3.103448275862069e-05, + "loss": 0.014, + "step": 1254 + }, + { + "epoch": 21.655172413793103, + "grad_norm": 0.47104957699775696, + "learning_rate": 3.090676883780332e-05, + "loss": 0.0057, + "step": 1256 + }, + { + "epoch": 21.689655172413794, + "grad_norm": 2.721670150756836, + "learning_rate": 3.0779054916985955e-05, + "loss": 0.0277, + "step": 1258 + }, + { + "epoch": 21.724137931034484, + "grad_norm": 0.09523604065179825, + "learning_rate": 3.065134099616858e-05, + "loss": 0.0015, + "step": 1260 + }, + { + "epoch": 21.75862068965517, + "grad_norm": 4.91738748550415, + "learning_rate": 3.052362707535121e-05, + "loss": 0.0663, + "step": 1262 + }, + { + "epoch": 21.79310344827586, + "grad_norm": 0.12091673910617828, + "learning_rate": 3.0395913154533843e-05, + "loss": 0.0024, + "step": 1264 + }, + { + "epoch": 21.82758620689655, + "grad_norm": 11.164044380187988, + "learning_rate": 3.0268199233716475e-05, + "loss": 0.0475, + "step": 1266 + }, + { + "epoch": 21.862068965517242, + "grad_norm": 0.018948553130030632, + "learning_rate": 3.0140485312899107e-05, + "loss": 0.0597, + "step": 1268 + }, + { + "epoch": 21.896551724137932, + "grad_norm": 4.384052753448486, + "learning_rate": 3.0012771392081738e-05, + "loss": 0.1747, + "step": 1270 + }, + { + "epoch": 21.93103448275862, + "grad_norm": 8.05239486694336, + "learning_rate": 2.988505747126437e-05, + "loss": 0.0828, + "step": 1272 + }, + { + "epoch": 21.96551724137931, + "grad_norm": 0.07449853420257568, + "learning_rate": 2.9757343550446998e-05, + "loss": 0.0028, + "step": 1274 + }, + { + "epoch": 22.0, + "grad_norm": 6.600012302398682, + "learning_rate": 2.962962962962963e-05, + "loss": 0.0353, + "step": 1276 + }, + { + "epoch": 22.0, + "eval_accuracy": 0.7045454545454546, + "eval_f1_macro": 0.6136674105378829, + "eval_f1_micro": 0.7045454545454546, + "eval_f1_weighted": 0.6918989425351217, + "eval_loss": 1.448930263519287, + "eval_precision_macro": 0.6146465063758297, + "eval_precision_micro": 0.7045454545454546, + "eval_precision_weighted": 0.6908581184896975, + "eval_recall_macro": 0.6242167919799498, + "eval_recall_micro": 0.7045454545454546, + "eval_recall_weighted": 0.7045454545454546, + "eval_runtime": 2.3557, + "eval_samples_per_second": 56.034, + "eval_steps_per_second": 7.217, + "step": 1276 + }, + { + "epoch": 22.03448275862069, + "grad_norm": 0.022473517805337906, + "learning_rate": 2.950191570881226e-05, + "loss": 0.0027, + "step": 1278 + }, + { + "epoch": 22.06896551724138, + "grad_norm": 0.06254889816045761, + "learning_rate": 2.9374201787994893e-05, + "loss": 0.0087, + "step": 1280 + }, + { + "epoch": 22.103448275862068, + "grad_norm": 0.07748962938785553, + "learning_rate": 2.9246487867177525e-05, + "loss": 0.0042, + "step": 1282 + }, + { + "epoch": 22.137931034482758, + "grad_norm": 0.07348134368658066, + "learning_rate": 2.9118773946360157e-05, + "loss": 0.0018, + "step": 1284 + }, + { + "epoch": 22.17241379310345, + "grad_norm": 5.811524868011475, + "learning_rate": 2.899106002554279e-05, + "loss": 0.0482, + "step": 1286 + }, + { + "epoch": 22.20689655172414, + "grad_norm": 0.06995466351509094, + "learning_rate": 2.8863346104725413e-05, + "loss": 0.002, + "step": 1288 + }, + { + "epoch": 22.24137931034483, + "grad_norm": 0.033994317054748535, + "learning_rate": 2.8735632183908045e-05, + "loss": 0.0219, + "step": 1290 + }, + { + "epoch": 22.275862068965516, + "grad_norm": 0.03432360664010048, + "learning_rate": 2.8607918263090677e-05, + "loss": 0.0009, + "step": 1292 + }, + { + "epoch": 22.310344827586206, + "grad_norm": 0.01812071166932583, + "learning_rate": 2.8480204342273308e-05, + "loss": 0.0015, + "step": 1294 + }, + { + "epoch": 22.344827586206897, + "grad_norm": 1.0397570133209229, + "learning_rate": 2.835249042145594e-05, + "loss": 0.0078, + "step": 1296 + }, + { + "epoch": 22.379310344827587, + "grad_norm": 0.5670585036277771, + "learning_rate": 2.822477650063857e-05, + "loss": 0.0359, + "step": 1298 + }, + { + "epoch": 22.413793103448278, + "grad_norm": 0.03518284112215042, + "learning_rate": 2.8097062579821203e-05, + "loss": 0.001, + "step": 1300 + }, + { + "epoch": 22.448275862068964, + "grad_norm": 0.3894251585006714, + "learning_rate": 2.796934865900383e-05, + "loss": 0.0021, + "step": 1302 + }, + { + "epoch": 22.482758620689655, + "grad_norm": 0.12059098482131958, + "learning_rate": 2.7841634738186463e-05, + "loss": 0.0014, + "step": 1304 + }, + { + "epoch": 22.517241379310345, + "grad_norm": 3.6728742122650146, + "learning_rate": 2.7713920817369095e-05, + "loss": 0.0199, + "step": 1306 + }, + { + "epoch": 22.551724137931036, + "grad_norm": 0.029478440061211586, + "learning_rate": 2.7586206896551727e-05, + "loss": 0.0009, + "step": 1308 + }, + { + "epoch": 22.586206896551722, + "grad_norm": 0.31572234630584717, + "learning_rate": 2.745849297573436e-05, + "loss": 0.004, + "step": 1310 + }, + { + "epoch": 22.620689655172413, + "grad_norm": 0.04008958116173744, + "learning_rate": 2.733077905491699e-05, + "loss": 0.0017, + "step": 1312 + }, + { + "epoch": 22.655172413793103, + "grad_norm": 1.0323200225830078, + "learning_rate": 2.720306513409962e-05, + "loss": 0.0077, + "step": 1314 + }, + { + "epoch": 22.689655172413794, + "grad_norm": 0.032519642263650894, + "learning_rate": 2.7075351213282253e-05, + "loss": 0.0027, + "step": 1316 + }, + { + "epoch": 22.724137931034484, + "grad_norm": 2.1410605907440186, + "learning_rate": 2.6947637292464878e-05, + "loss": 0.0125, + "step": 1318 + }, + { + "epoch": 22.75862068965517, + "grad_norm": 0.016050921753048897, + "learning_rate": 2.681992337164751e-05, + "loss": 0.0331, + "step": 1320 + }, + { + "epoch": 22.79310344827586, + "grad_norm": 0.1428113877773285, + "learning_rate": 2.669220945083014e-05, + "loss": 0.0049, + "step": 1322 + }, + { + "epoch": 22.82758620689655, + "grad_norm": 0.1641322523355484, + "learning_rate": 2.6564495530012773e-05, + "loss": 0.002, + "step": 1324 + }, + { + "epoch": 22.862068965517242, + "grad_norm": 7.986566066741943, + "learning_rate": 2.6436781609195405e-05, + "loss": 0.0351, + "step": 1326 + }, + { + "epoch": 22.896551724137932, + "grad_norm": 0.08804414421319962, + "learning_rate": 2.6309067688378037e-05, + "loss": 0.0013, + "step": 1328 + }, + { + "epoch": 22.93103448275862, + "grad_norm": 7.736631870269775, + "learning_rate": 2.618135376756067e-05, + "loss": 0.0779, + "step": 1330 + }, + { + "epoch": 22.96551724137931, + "grad_norm": 0.015013402327895164, + "learning_rate": 2.6053639846743293e-05, + "loss": 0.0827, + "step": 1332 + }, + { + "epoch": 23.0, + "grad_norm": 0.01156703382730484, + "learning_rate": 2.5925925925925925e-05, + "loss": 0.001, + "step": 1334 + }, + { + "epoch": 23.0, + "eval_accuracy": 0.6666666666666666, + "eval_f1_macro": 0.5714664821494502, + "eval_f1_micro": 0.6666666666666666, + "eval_f1_weighted": 0.6541314709388933, + "eval_loss": 1.4781274795532227, + "eval_precision_macro": 0.5657452925012241, + "eval_precision_micro": 0.6666666666666666, + "eval_precision_weighted": 0.6448584800964208, + "eval_recall_macro": 0.580545112781955, + "eval_recall_micro": 0.6666666666666666, + "eval_recall_weighted": 0.6666666666666666, + "eval_runtime": 1.954, + "eval_samples_per_second": 67.553, + "eval_steps_per_second": 8.7, + "step": 1334 + }, + { + "epoch": 23.03448275862069, + "grad_norm": 0.006409560330212116, + "learning_rate": 2.5798212005108557e-05, + "loss": 0.0018, + "step": 1336 + }, + { + "epoch": 23.06896551724138, + "grad_norm": 2.2381739616394043, + "learning_rate": 2.5670498084291188e-05, + "loss": 0.0467, + "step": 1338 + }, + { + "epoch": 23.103448275862068, + "grad_norm": 0.5874573588371277, + "learning_rate": 2.554278416347382e-05, + "loss": 0.0048, + "step": 1340 + }, + { + "epoch": 23.137931034482758, + "grad_norm": 22.487873077392578, + "learning_rate": 2.541507024265645e-05, + "loss": 0.1, + "step": 1342 + }, + { + "epoch": 23.17241379310345, + "grad_norm": 0.020600179210305214, + "learning_rate": 2.5287356321839083e-05, + "loss": 0.025, + "step": 1344 + }, + { + "epoch": 23.20689655172414, + "grad_norm": 0.17713558673858643, + "learning_rate": 2.515964240102171e-05, + "loss": 0.0037, + "step": 1346 + }, + { + "epoch": 23.24137931034483, + "grad_norm": 0.19636306166648865, + "learning_rate": 2.5031928480204343e-05, + "loss": 0.0024, + "step": 1348 + }, + { + "epoch": 23.275862068965516, + "grad_norm": 0.5884350538253784, + "learning_rate": 2.4904214559386975e-05, + "loss": 0.0041, + "step": 1350 + }, + { + "epoch": 23.310344827586206, + "grad_norm": 0.07146797329187393, + "learning_rate": 2.4776500638569607e-05, + "loss": 0.0012, + "step": 1352 + }, + { + "epoch": 23.344827586206897, + "grad_norm": 0.11290651559829712, + "learning_rate": 2.464878671775224e-05, + "loss": 0.008, + "step": 1354 + }, + { + "epoch": 23.379310344827587, + "grad_norm": 0.09537604451179504, + "learning_rate": 2.4521072796934867e-05, + "loss": 0.0339, + "step": 1356 + }, + { + "epoch": 23.413793103448278, + "grad_norm": 0.27110394835472107, + "learning_rate": 2.4393358876117498e-05, + "loss": 0.0021, + "step": 1358 + }, + { + "epoch": 23.448275862068964, + "grad_norm": 0.2351570874452591, + "learning_rate": 2.426564495530013e-05, + "loss": 0.0015, + "step": 1360 + }, + { + "epoch": 23.482758620689655, + "grad_norm": 0.018089979887008667, + "learning_rate": 2.413793103448276e-05, + "loss": 0.0023, + "step": 1362 + }, + { + "epoch": 23.517241379310345, + "grad_norm": 0.12375891953706741, + "learning_rate": 2.401021711366539e-05, + "loss": 0.0035, + "step": 1364 + }, + { + "epoch": 23.551724137931036, + "grad_norm": 0.2663002908229828, + "learning_rate": 2.388250319284802e-05, + "loss": 0.0048, + "step": 1366 + }, + { + "epoch": 23.586206896551722, + "grad_norm": 0.031247558072209358, + "learning_rate": 2.3754789272030653e-05, + "loss": 0.092, + "step": 1368 + }, + { + "epoch": 23.620689655172413, + "grad_norm": 0.024893470108509064, + "learning_rate": 2.3627075351213285e-05, + "loss": 0.0008, + "step": 1370 + }, + { + "epoch": 23.655172413793103, + "grad_norm": 0.040011290460824966, + "learning_rate": 2.3499361430395913e-05, + "loss": 0.0017, + "step": 1372 + }, + { + "epoch": 23.689655172413794, + "grad_norm": 0.6574517488479614, + "learning_rate": 2.3371647509578545e-05, + "loss": 0.024, + "step": 1374 + }, + { + "epoch": 23.724137931034484, + "grad_norm": 0.02834693342447281, + "learning_rate": 2.3243933588761177e-05, + "loss": 0.0006, + "step": 1376 + }, + { + "epoch": 23.75862068965517, + "grad_norm": 1.0888547897338867, + "learning_rate": 2.3116219667943805e-05, + "loss": 0.006, + "step": 1378 + }, + { + "epoch": 23.79310344827586, + "grad_norm": 0.03572368249297142, + "learning_rate": 2.2988505747126437e-05, + "loss": 0.0028, + "step": 1380 + }, + { + "epoch": 23.82758620689655, + "grad_norm": 5.217251777648926, + "learning_rate": 2.2860791826309068e-05, + "loss": 0.1481, + "step": 1382 + }, + { + "epoch": 23.862068965517242, + "grad_norm": 0.4067634046077728, + "learning_rate": 2.27330779054917e-05, + "loss": 0.0045, + "step": 1384 + }, + { + "epoch": 23.896551724137932, + "grad_norm": 0.1273973435163498, + "learning_rate": 2.2605363984674328e-05, + "loss": 0.0013, + "step": 1386 + }, + { + "epoch": 23.93103448275862, + "grad_norm": 0.3500019907951355, + "learning_rate": 2.247765006385696e-05, + "loss": 0.0023, + "step": 1388 + }, + { + "epoch": 23.96551724137931, + "grad_norm": 2.0916292667388916, + "learning_rate": 2.234993614303959e-05, + "loss": 0.0074, + "step": 1390 + }, + { + "epoch": 24.0, + "grad_norm": 0.03820287436246872, + "learning_rate": 2.2222222222222223e-05, + "loss": 0.0007, + "step": 1392 + }, + { + "epoch": 24.0, + "eval_accuracy": 0.6590909090909091, + "eval_f1_macro": 0.5712607596935955, + "eval_f1_micro": 0.6590909090909091, + "eval_f1_weighted": 0.6510971028694909, + "eval_loss": 1.632594347000122, + "eval_precision_macro": 0.5870793772054277, + "eval_precision_micro": 0.6590909090909091, + "eval_precision_weighted": 0.664782083632351, + "eval_recall_macro": 0.5786027568922305, + "eval_recall_micro": 0.6590909090909091, + "eval_recall_weighted": 0.6590909090909091, + "eval_runtime": 1.9749, + "eval_samples_per_second": 66.84, + "eval_steps_per_second": 8.608, + "step": 1392 + }, + { + "epoch": 24.03448275862069, + "grad_norm": 0.07503662258386612, + "learning_rate": 2.2094508301404855e-05, + "loss": 0.0244, + "step": 1394 + }, + { + "epoch": 24.06896551724138, + "grad_norm": 27.82219123840332, + "learning_rate": 2.1966794380587487e-05, + "loss": 0.196, + "step": 1396 + }, + { + "epoch": 24.103448275862068, + "grad_norm": 0.029466254636645317, + "learning_rate": 2.183908045977012e-05, + "loss": 0.0007, + "step": 1398 + }, + { + "epoch": 24.137931034482758, + "grad_norm": 0.15405841171741486, + "learning_rate": 2.1711366538952747e-05, + "loss": 0.0015, + "step": 1400 + }, + { + "epoch": 24.17241379310345, + "grad_norm": 2.168177366256714, + "learning_rate": 2.1583652618135378e-05, + "loss": 0.0114, + "step": 1402 + }, + { + "epoch": 24.20689655172414, + "grad_norm": 1.049078345298767, + "learning_rate": 2.145593869731801e-05, + "loss": 0.0039, + "step": 1404 + }, + { + "epoch": 24.24137931034483, + "grad_norm": 0.10657618939876556, + "learning_rate": 2.1328224776500638e-05, + "loss": 0.0016, + "step": 1406 + }, + { + "epoch": 24.275862068965516, + "grad_norm": 2.1004295349121094, + "learning_rate": 2.120051085568327e-05, + "loss": 0.0427, + "step": 1408 + }, + { + "epoch": 24.310344827586206, + "grad_norm": 0.026559889316558838, + "learning_rate": 2.10727969348659e-05, + "loss": 0.0008, + "step": 1410 + }, + { + "epoch": 24.344827586206897, + "grad_norm": 0.2494554966688156, + "learning_rate": 2.0945083014048533e-05, + "loss": 0.0021, + "step": 1412 + }, + { + "epoch": 24.379310344827587, + "grad_norm": 0.5512819290161133, + "learning_rate": 2.081736909323116e-05, + "loss": 0.0093, + "step": 1414 + }, + { + "epoch": 24.413793103448278, + "grad_norm": 0.032853931188583374, + "learning_rate": 2.0689655172413793e-05, + "loss": 0.0006, + "step": 1416 + }, + { + "epoch": 24.448275862068964, + "grad_norm": 0.29787227511405945, + "learning_rate": 2.0561941251596425e-05, + "loss": 0.0037, + "step": 1418 + }, + { + "epoch": 24.482758620689655, + "grad_norm": 0.22131195664405823, + "learning_rate": 2.0434227330779057e-05, + "loss": 0.0086, + "step": 1420 + }, + { + "epoch": 24.517241379310345, + "grad_norm": 0.15168803930282593, + "learning_rate": 2.0306513409961685e-05, + "loss": 0.0017, + "step": 1422 + }, + { + "epoch": 24.551724137931036, + "grad_norm": 0.006662359926849604, + "learning_rate": 2.0178799489144317e-05, + "loss": 0.1826, + "step": 1424 + }, + { + "epoch": 24.586206896551722, + "grad_norm": 0.008878910914063454, + "learning_rate": 2.0051085568326948e-05, + "loss": 0.0096, + "step": 1426 + }, + { + "epoch": 24.620689655172413, + "grad_norm": 0.09360943734645844, + "learning_rate": 1.992337164750958e-05, + "loss": 0.0015, + "step": 1428 + }, + { + "epoch": 24.655172413793103, + "grad_norm": 0.13396118581295013, + "learning_rate": 1.979565772669221e-05, + "loss": 0.3838, + "step": 1430 + }, + { + "epoch": 24.689655172413794, + "grad_norm": 0.010010026395320892, + "learning_rate": 1.9667943805874843e-05, + "loss": 0.0005, + "step": 1432 + }, + { + "epoch": 24.724137931034484, + "grad_norm": 1.2189490795135498, + "learning_rate": 1.9540229885057475e-05, + "loss": 0.0052, + "step": 1434 + }, + { + "epoch": 24.75862068965517, + "grad_norm": 0.0254853293299675, + "learning_rate": 1.9412515964240103e-05, + "loss": 0.007, + "step": 1436 + }, + { + "epoch": 24.79310344827586, + "grad_norm": 5.6452765464782715, + "learning_rate": 1.9284802043422735e-05, + "loss": 0.0186, + "step": 1438 + }, + { + "epoch": 24.82758620689655, + "grad_norm": 0.12432337552309036, + "learning_rate": 1.9157088122605367e-05, + "loss": 0.0012, + "step": 1440 + }, + { + "epoch": 24.862068965517242, + "grad_norm": 0.06803122162818909, + "learning_rate": 1.9029374201787995e-05, + "loss": 0.001, + "step": 1442 + }, + { + "epoch": 24.896551724137932, + "grad_norm": 0.07736227661371231, + "learning_rate": 1.8901660280970627e-05, + "loss": 0.0007, + "step": 1444 + }, + { + "epoch": 24.93103448275862, + "grad_norm": 0.140510693192482, + "learning_rate": 1.8773946360153258e-05, + "loss": 0.0032, + "step": 1446 + }, + { + "epoch": 24.96551724137931, + "grad_norm": 0.1517113447189331, + "learning_rate": 1.864623243933589e-05, + "loss": 0.0142, + "step": 1448 + }, + { + "epoch": 25.0, + "grad_norm": 1.449758768081665, + "learning_rate": 1.8518518518518518e-05, + "loss": 0.0084, + "step": 1450 + }, + { + "epoch": 25.0, + "eval_accuracy": 0.6590909090909091, + "eval_f1_macro": 0.5684048691621123, + "eval_f1_micro": 0.6590909090909091, + "eval_f1_weighted": 0.6569451587148032, + "eval_loss": 1.5855995416641235, + "eval_precision_macro": 0.5662025575780552, + "eval_precision_micro": 0.6590909090909091, + "eval_precision_weighted": 0.6671501716534084, + "eval_recall_macro": 0.5802474937343358, + "eval_recall_micro": 0.6590909090909091, + "eval_recall_weighted": 0.6590909090909091, + "eval_runtime": 2.0013, + "eval_samples_per_second": 65.957, + "eval_steps_per_second": 8.495, + "step": 1450 + }, + { + "epoch": 25.03448275862069, + "grad_norm": 3.8146300315856934, + "learning_rate": 1.839080459770115e-05, + "loss": 0.116, + "step": 1452 + }, + { + "epoch": 25.06896551724138, + "grad_norm": 0.013178830966353416, + "learning_rate": 1.826309067688378e-05, + "loss": 0.0004, + "step": 1454 + }, + { + "epoch": 25.103448275862068, + "grad_norm": 0.04289768263697624, + "learning_rate": 1.8135376756066413e-05, + "loss": 0.001, + "step": 1456 + }, + { + "epoch": 25.137931034482758, + "grad_norm": 0.3486016094684601, + "learning_rate": 1.800766283524904e-05, + "loss": 0.0018, + "step": 1458 + }, + { + "epoch": 25.17241379310345, + "grad_norm": 0.101795494556427, + "learning_rate": 1.7879948914431673e-05, + "loss": 0.0487, + "step": 1460 + }, + { + "epoch": 25.20689655172414, + "grad_norm": 0.008328469470143318, + "learning_rate": 1.7752234993614305e-05, + "loss": 0.0003, + "step": 1462 + }, + { + "epoch": 25.24137931034483, + "grad_norm": 0.022965148091316223, + "learning_rate": 1.7624521072796933e-05, + "loss": 0.0007, + "step": 1464 + }, + { + "epoch": 25.275862068965516, + "grad_norm": 0.05658239498734474, + "learning_rate": 1.7496807151979565e-05, + "loss": 0.0012, + "step": 1466 + }, + { + "epoch": 25.310344827586206, + "grad_norm": 0.4014154374599457, + "learning_rate": 1.7369093231162197e-05, + "loss": 0.0046, + "step": 1468 + }, + { + "epoch": 25.344827586206897, + "grad_norm": 5.419172763824463, + "learning_rate": 1.7241379310344828e-05, + "loss": 0.0219, + "step": 1470 + }, + { + "epoch": 25.379310344827587, + "grad_norm": 0.022299258038401604, + "learning_rate": 1.711366538952746e-05, + "loss": 0.0007, + "step": 1472 + }, + { + "epoch": 25.413793103448278, + "grad_norm": 0.8597404360771179, + "learning_rate": 1.698595146871009e-05, + "loss": 0.0042, + "step": 1474 + }, + { + "epoch": 25.448275862068964, + "grad_norm": 0.056232478469610214, + "learning_rate": 1.6858237547892723e-05, + "loss": 0.0007, + "step": 1476 + }, + { + "epoch": 25.482758620689655, + "grad_norm": 0.8406282663345337, + "learning_rate": 1.673052362707535e-05, + "loss": 0.0036, + "step": 1478 + }, + { + "epoch": 25.517241379310345, + "grad_norm": 0.03363404422998428, + "learning_rate": 1.6602809706257983e-05, + "loss": 0.0014, + "step": 1480 + }, + { + "epoch": 25.551724137931036, + "grad_norm": 0.015727031975984573, + "learning_rate": 1.6475095785440615e-05, + "loss": 0.0008, + "step": 1482 + }, + { + "epoch": 25.586206896551722, + "grad_norm": 0.03139304369688034, + "learning_rate": 1.6347381864623247e-05, + "loss": 0.0017, + "step": 1484 + }, + { + "epoch": 25.620689655172413, + "grad_norm": 0.02697753719985485, + "learning_rate": 1.6219667943805875e-05, + "loss": 0.0008, + "step": 1486 + }, + { + "epoch": 25.655172413793103, + "grad_norm": 0.004935800563544035, + "learning_rate": 1.6091954022988507e-05, + "loss": 0.0006, + "step": 1488 + }, + { + "epoch": 25.689655172413794, + "grad_norm": 1.9771162271499634, + "learning_rate": 1.596424010217114e-05, + "loss": 0.0139, + "step": 1490 + }, + { + "epoch": 25.724137931034484, + "grad_norm": 0.014661573804914951, + "learning_rate": 1.583652618135377e-05, + "loss": 0.0004, + "step": 1492 + }, + { + "epoch": 25.75862068965517, + "grad_norm": 2.9789557456970215, + "learning_rate": 1.5708812260536398e-05, + "loss": 0.0784, + "step": 1494 + }, + { + "epoch": 25.79310344827586, + "grad_norm": 0.05912244692444801, + "learning_rate": 1.558109833971903e-05, + "loss": 0.002, + "step": 1496 + }, + { + "epoch": 25.82758620689655, + "grad_norm": 0.21335746347904205, + "learning_rate": 1.545338441890166e-05, + "loss": 0.002, + "step": 1498 + }, + { + "epoch": 25.862068965517242, + "grad_norm": 0.01072037685662508, + "learning_rate": 1.532567049808429e-05, + "loss": 0.0005, + "step": 1500 + }, + { + "epoch": 25.896551724137932, + "grad_norm": 0.04208362475037575, + "learning_rate": 1.5197956577266922e-05, + "loss": 0.0007, + "step": 1502 + }, + { + "epoch": 25.93103448275862, + "grad_norm": 0.026036394760012627, + "learning_rate": 1.5070242656449553e-05, + "loss": 0.0006, + "step": 1504 + }, + { + "epoch": 25.96551724137931, + "grad_norm": 0.08341384679079056, + "learning_rate": 1.4942528735632185e-05, + "loss": 0.0012, + "step": 1506 + }, + { + "epoch": 26.0, + "grad_norm": 0.0347595140337944, + "learning_rate": 1.4814814814814815e-05, + "loss": 0.0008, + "step": 1508 + }, + { + "epoch": 26.0, + "eval_accuracy": 0.6818181818181818, + "eval_f1_macro": 0.5826228878648233, + "eval_f1_micro": 0.6818181818181818, + "eval_f1_weighted": 0.6675219941348973, + "eval_loss": 1.5798661708831787, + "eval_precision_macro": 0.5849358974358975, + "eval_precision_micro": 0.6818181818181818, + "eval_precision_weighted": 0.6631993006993008, + "eval_recall_macro": 0.5884241854636592, + "eval_recall_micro": 0.6818181818181818, + "eval_recall_weighted": 0.6818181818181818, + "eval_runtime": 1.9704, + "eval_samples_per_second": 66.992, + "eval_steps_per_second": 8.628, + "step": 1508 + }, + { + "epoch": 26.03448275862069, + "grad_norm": 0.10061348974704742, + "learning_rate": 1.4687100893997447e-05, + "loss": 0.0221, + "step": 1510 + }, + { + "epoch": 26.06896551724138, + "grad_norm": 0.005162129178643227, + "learning_rate": 1.4559386973180078e-05, + "loss": 0.0003, + "step": 1512 + }, + { + "epoch": 26.103448275862068, + "grad_norm": 0.02409091778099537, + "learning_rate": 1.4431673052362707e-05, + "loss": 0.0004, + "step": 1514 + }, + { + "epoch": 26.137931034482758, + "grad_norm": 0.02003251574933529, + "learning_rate": 1.4303959131545338e-05, + "loss": 0.0009, + "step": 1516 + }, + { + "epoch": 26.17241379310345, + "grad_norm": 0.11420497298240662, + "learning_rate": 1.417624521072797e-05, + "loss": 0.0011, + "step": 1518 + }, + { + "epoch": 26.20689655172414, + "grad_norm": 0.03824389725923538, + "learning_rate": 1.4048531289910602e-05, + "loss": 0.0008, + "step": 1520 + }, + { + "epoch": 26.24137931034483, + "grad_norm": 0.00969379860907793, + "learning_rate": 1.3920817369093232e-05, + "loss": 0.0003, + "step": 1522 + }, + { + "epoch": 26.275862068965516, + "grad_norm": 1.241509199142456, + "learning_rate": 1.3793103448275863e-05, + "loss": 0.0038, + "step": 1524 + }, + { + "epoch": 26.310344827586206, + "grad_norm": 0.011483917012810707, + "learning_rate": 1.3665389527458495e-05, + "loss": 0.0011, + "step": 1526 + }, + { + "epoch": 26.344827586206897, + "grad_norm": 0.02756449580192566, + "learning_rate": 1.3537675606641127e-05, + "loss": 0.0007, + "step": 1528 + }, + { + "epoch": 26.379310344827587, + "grad_norm": 0.01528858207166195, + "learning_rate": 1.3409961685823755e-05, + "loss": 0.0003, + "step": 1530 + }, + { + "epoch": 26.413793103448278, + "grad_norm": 0.01113022118806839, + "learning_rate": 1.3282247765006387e-05, + "loss": 0.0005, + "step": 1532 + }, + { + "epoch": 26.448275862068964, + "grad_norm": 18.05221176147461, + "learning_rate": 1.3154533844189018e-05, + "loss": 0.1686, + "step": 1534 + }, + { + "epoch": 26.482758620689655, + "grad_norm": 0.009326491504907608, + "learning_rate": 1.3026819923371647e-05, + "loss": 0.0008, + "step": 1536 + }, + { + "epoch": 26.517241379310345, + "grad_norm": 0.04548182338476181, + "learning_rate": 1.2899106002554278e-05, + "loss": 0.0007, + "step": 1538 + }, + { + "epoch": 26.551724137931036, + "grad_norm": 0.008745373226702213, + "learning_rate": 1.277139208173691e-05, + "loss": 0.0004, + "step": 1540 + }, + { + "epoch": 26.586206896551722, + "grad_norm": 0.24436865746974945, + "learning_rate": 1.2643678160919542e-05, + "loss": 0.0021, + "step": 1542 + }, + { + "epoch": 26.620689655172413, + "grad_norm": 0.01390091422945261, + "learning_rate": 1.2515964240102172e-05, + "loss": 0.0008, + "step": 1544 + }, + { + "epoch": 26.655172413793103, + "grad_norm": 0.016228720545768738, + "learning_rate": 1.2388250319284803e-05, + "loss": 0.0009, + "step": 1546 + }, + { + "epoch": 26.689655172413794, + "grad_norm": 1.8321493864059448, + "learning_rate": 1.2260536398467433e-05, + "loss": 0.0245, + "step": 1548 + }, + { + "epoch": 26.724137931034484, + "grad_norm": 0.11651375144720078, + "learning_rate": 1.2132822477650065e-05, + "loss": 0.0012, + "step": 1550 + }, + { + "epoch": 26.75862068965517, + "grad_norm": 0.26734495162963867, + "learning_rate": 1.2005108556832695e-05, + "loss": 0.0306, + "step": 1552 + }, + { + "epoch": 26.79310344827586, + "grad_norm": 1.002111792564392, + "learning_rate": 1.1877394636015327e-05, + "loss": 0.0049, + "step": 1554 + }, + { + "epoch": 26.82758620689655, + "grad_norm": 0.014619125053286552, + "learning_rate": 1.1749680715197957e-05, + "loss": 0.0012, + "step": 1556 + }, + { + "epoch": 26.862068965517242, + "grad_norm": 0.18940378725528717, + "learning_rate": 1.1621966794380588e-05, + "loss": 0.0013, + "step": 1558 + }, + { + "epoch": 26.896551724137932, + "grad_norm": 0.002744639990851283, + "learning_rate": 1.1494252873563218e-05, + "loss": 0.0003, + "step": 1560 + }, + { + "epoch": 26.93103448275862, + "grad_norm": 0.14458243548870087, + "learning_rate": 1.136653895274585e-05, + "loss": 0.0017, + "step": 1562 + }, + { + "epoch": 26.96551724137931, + "grad_norm": 0.44180047512054443, + "learning_rate": 1.123882503192848e-05, + "loss": 0.0033, + "step": 1564 + }, + { + "epoch": 27.0, + "grad_norm": 1.8280788660049438, + "learning_rate": 1.1111111111111112e-05, + "loss": 0.0053, + "step": 1566 + }, + { + "epoch": 27.0, + "eval_accuracy": 0.6666666666666666, + "eval_f1_macro": 0.5719249734426743, + "eval_f1_micro": 0.6666666666666666, + "eval_f1_weighted": 0.6555785069889042, + "eval_loss": 1.5308274030685425, + "eval_precision_macro": 0.5667354028698567, + "eval_precision_micro": 0.6666666666666666, + "eval_precision_weighted": 0.652369328839917, + "eval_recall_macro": 0.5842575187969924, + "eval_recall_micro": 0.6666666666666666, + "eval_recall_weighted": 0.6666666666666666, + "eval_runtime": 1.9794, + "eval_samples_per_second": 66.687, + "eval_steps_per_second": 8.588, + "step": 1566 + }, + { + "epoch": 27.03448275862069, + "grad_norm": 0.028027068823575974, + "learning_rate": 1.0983397190293743e-05, + "loss": 0.0007, + "step": 1568 + }, + { + "epoch": 27.06896551724138, + "grad_norm": 0.017265431582927704, + "learning_rate": 1.0855683269476373e-05, + "loss": 0.0006, + "step": 1570 + }, + { + "epoch": 27.103448275862068, + "grad_norm": 0.004702151753008366, + "learning_rate": 1.0727969348659005e-05, + "loss": 0.0078, + "step": 1572 + }, + { + "epoch": 27.137931034482758, + "grad_norm": 1.1751693487167358, + "learning_rate": 1.0600255427841635e-05, + "loss": 0.0102, + "step": 1574 + }, + { + "epoch": 27.17241379310345, + "grad_norm": 0.042279455810785294, + "learning_rate": 1.0472541507024267e-05, + "loss": 0.001, + "step": 1576 + }, + { + "epoch": 27.20689655172414, + "grad_norm": 0.013697315938770771, + "learning_rate": 1.0344827586206897e-05, + "loss": 0.0005, + "step": 1578 + }, + { + "epoch": 27.24137931034483, + "grad_norm": 0.1087651401758194, + "learning_rate": 1.0217113665389528e-05, + "loss": 0.0009, + "step": 1580 + }, + { + "epoch": 27.275862068965516, + "grad_norm": 0.040765080600976944, + "learning_rate": 1.0089399744572158e-05, + "loss": 0.0015, + "step": 1582 + }, + { + "epoch": 27.310344827586206, + "grad_norm": 0.15862275660037994, + "learning_rate": 9.96168582375479e-06, + "loss": 0.0029, + "step": 1584 + }, + { + "epoch": 27.344827586206897, + "grad_norm": 0.014304181560873985, + "learning_rate": 9.833971902937422e-06, + "loss": 0.045, + "step": 1586 + }, + { + "epoch": 27.379310344827587, + "grad_norm": 0.4837823808193207, + "learning_rate": 9.706257982120052e-06, + "loss": 0.003, + "step": 1588 + }, + { + "epoch": 27.413793103448278, + "grad_norm": 0.00520712835714221, + "learning_rate": 9.578544061302683e-06, + "loss": 0.0008, + "step": 1590 + }, + { + "epoch": 27.448275862068964, + "grad_norm": 0.026446396484971046, + "learning_rate": 9.450830140485313e-06, + "loss": 0.0005, + "step": 1592 + }, + { + "epoch": 27.482758620689655, + "grad_norm": 0.021808737888932228, + "learning_rate": 9.323116219667945e-06, + "loss": 0.0008, + "step": 1594 + }, + { + "epoch": 27.517241379310345, + "grad_norm": 1.217366337776184, + "learning_rate": 9.195402298850575e-06, + "loss": 0.0255, + "step": 1596 + }, + { + "epoch": 27.551724137931036, + "grad_norm": 0.04630236327648163, + "learning_rate": 9.067688378033207e-06, + "loss": 0.0011, + "step": 1598 + }, + { + "epoch": 27.586206896551722, + "grad_norm": 0.029485022649168968, + "learning_rate": 8.939974457215837e-06, + "loss": 0.0004, + "step": 1600 + }, + { + "epoch": 27.620689655172413, + "grad_norm": 0.6011648178100586, + "learning_rate": 8.812260536398467e-06, + "loss": 0.0031, + "step": 1602 + }, + { + "epoch": 27.655172413793103, + "grad_norm": 0.016739701852202415, + "learning_rate": 8.684546615581098e-06, + "loss": 0.0005, + "step": 1604 + }, + { + "epoch": 27.689655172413794, + "grad_norm": 0.07989770919084549, + "learning_rate": 8.55683269476373e-06, + "loss": 0.0007, + "step": 1606 + }, + { + "epoch": 27.724137931034484, + "grad_norm": 0.009848535992205143, + "learning_rate": 8.429118773946362e-06, + "loss": 0.0004, + "step": 1608 + }, + { + "epoch": 27.75862068965517, + "grad_norm": 0.007213211618363857, + "learning_rate": 8.301404853128992e-06, + "loss": 0.0007, + "step": 1610 + }, + { + "epoch": 27.79310344827586, + "grad_norm": 0.08470471948385239, + "learning_rate": 8.173690932311623e-06, + "loss": 0.0017, + "step": 1612 + }, + { + "epoch": 27.82758620689655, + "grad_norm": 0.11105465888977051, + "learning_rate": 8.045977011494253e-06, + "loss": 0.001, + "step": 1614 + }, + { + "epoch": 27.862068965517242, + "grad_norm": 0.38511380553245544, + "learning_rate": 7.918263090676885e-06, + "loss": 0.0039, + "step": 1616 + }, + { + "epoch": 27.896551724137932, + "grad_norm": 0.07350783795118332, + "learning_rate": 7.790549169859515e-06, + "loss": 0.0011, + "step": 1618 + }, + { + "epoch": 27.93103448275862, + "grad_norm": 0.0037571904249489307, + "learning_rate": 7.662835249042145e-06, + "loss": 0.0004, + "step": 1620 + }, + { + "epoch": 27.96551724137931, + "grad_norm": 0.283407986164093, + "learning_rate": 7.535121328224777e-06, + "loss": 0.0016, + "step": 1622 + }, + { + "epoch": 28.0, + "grad_norm": 0.025982772931456566, + "learning_rate": 7.4074074074074075e-06, + "loss": 0.0004, + "step": 1624 + }, + { + "epoch": 28.0, + "eval_accuracy": 0.6666666666666666, + "eval_f1_macro": 0.5732077187394248, + "eval_f1_micro": 0.6666666666666666, + "eval_f1_weighted": 0.6617322262037608, + "eval_loss": 1.5639015436172485, + "eval_precision_macro": 0.5684216325737007, + "eval_precision_micro": 0.6666666666666666, + "eval_precision_weighted": 0.667311206663744, + "eval_recall_macro": 0.5867167919799499, + "eval_recall_micro": 0.6666666666666666, + "eval_recall_weighted": 0.6666666666666666, + "eval_runtime": 2.0384, + "eval_samples_per_second": 64.757, + "eval_steps_per_second": 8.34, + "step": 1624 + }, + { + "epoch": 28.03448275862069, + "grad_norm": 0.23507054150104523, + "learning_rate": 7.279693486590039e-06, + "loss": 0.003, + "step": 1626 + }, + { + "epoch": 28.06896551724138, + "grad_norm": 0.0050230068154633045, + "learning_rate": 7.151979565772669e-06, + "loss": 0.0006, + "step": 1628 + }, + { + "epoch": 28.103448275862068, + "grad_norm": 0.1437883973121643, + "learning_rate": 7.024265644955301e-06, + "loss": 0.0018, + "step": 1630 + }, + { + "epoch": 28.137931034482758, + "grad_norm": 1.1382052898406982, + "learning_rate": 6.896551724137932e-06, + "loss": 0.0066, + "step": 1632 + }, + { + "epoch": 28.17241379310345, + "grad_norm": 0.00969479139894247, + "learning_rate": 6.768837803320563e-06, + "loss": 0.0004, + "step": 1634 + }, + { + "epoch": 28.20689655172414, + "grad_norm": 0.015258271247148514, + "learning_rate": 6.641123882503193e-06, + "loss": 0.0006, + "step": 1636 + }, + { + "epoch": 28.24137931034483, + "grad_norm": 0.057105790823698044, + "learning_rate": 6.513409961685823e-06, + "loss": 0.0007, + "step": 1638 + }, + { + "epoch": 28.275862068965516, + "grad_norm": 0.007530784234404564, + "learning_rate": 6.385696040868455e-06, + "loss": 0.0042, + "step": 1640 + }, + { + "epoch": 28.310344827586206, + "grad_norm": 0.05503424257040024, + "learning_rate": 6.257982120051086e-06, + "loss": 0.0009, + "step": 1642 + }, + { + "epoch": 28.344827586206897, + "grad_norm": 0.19767296314239502, + "learning_rate": 6.130268199233717e-06, + "loss": 0.1093, + "step": 1644 + }, + { + "epoch": 28.379310344827587, + "grad_norm": 0.023296566680073738, + "learning_rate": 6.0025542784163475e-06, + "loss": 0.001, + "step": 1646 + }, + { + "epoch": 28.413793103448278, + "grad_norm": 0.018203437328338623, + "learning_rate": 5.874840357598978e-06, + "loss": 0.0004, + "step": 1648 + }, + { + "epoch": 28.448275862068964, + "grad_norm": 0.042694464325904846, + "learning_rate": 5.747126436781609e-06, + "loss": 0.0007, + "step": 1650 + }, + { + "epoch": 28.482758620689655, + "grad_norm": 0.019331173971295357, + "learning_rate": 5.61941251596424e-06, + "loss": 0.0005, + "step": 1652 + }, + { + "epoch": 28.517241379310345, + "grad_norm": 0.03333083540201187, + "learning_rate": 5.491698595146872e-06, + "loss": 0.0011, + "step": 1654 + }, + { + "epoch": 28.551724137931036, + "grad_norm": 0.005839875899255276, + "learning_rate": 5.3639846743295025e-06, + "loss": 0.0006, + "step": 1656 + }, + { + "epoch": 28.586206896551722, + "grad_norm": 0.040753502398729324, + "learning_rate": 5.236270753512133e-06, + "loss": 0.0005, + "step": 1658 + }, + { + "epoch": 28.620689655172413, + "grad_norm": 0.0062615578062832355, + "learning_rate": 5.108556832694764e-06, + "loss": 0.0009, + "step": 1660 + }, + { + "epoch": 28.655172413793103, + "grad_norm": 0.024959621950984, + "learning_rate": 4.980842911877395e-06, + "loss": 0.0006, + "step": 1662 + }, + { + "epoch": 28.689655172413794, + "grad_norm": 0.006987102795392275, + "learning_rate": 4.853128991060026e-06, + "loss": 0.0007, + "step": 1664 + }, + { + "epoch": 28.724137931034484, + "grad_norm": 0.005346678197383881, + "learning_rate": 4.725415070242657e-06, + "loss": 0.0003, + "step": 1666 + }, + { + "epoch": 28.75862068965517, + "grad_norm": 0.08393517136573792, + "learning_rate": 4.5977011494252875e-06, + "loss": 0.0013, + "step": 1668 + }, + { + "epoch": 28.79310344827586, + "grad_norm": 0.01564914733171463, + "learning_rate": 4.469987228607918e-06, + "loss": 0.0206, + "step": 1670 + }, + { + "epoch": 28.82758620689655, + "grad_norm": 0.04356468468904495, + "learning_rate": 4.342273307790549e-06, + "loss": 0.0007, + "step": 1672 + }, + { + "epoch": 28.862068965517242, + "grad_norm": 0.056500744074583054, + "learning_rate": 4.214559386973181e-06, + "loss": 0.0011, + "step": 1674 + }, + { + "epoch": 28.896551724137932, + "grad_norm": 0.1394091099500656, + "learning_rate": 4.086845466155812e-06, + "loss": 0.0012, + "step": 1676 + }, + { + "epoch": 28.93103448275862, + "grad_norm": 8.025191307067871, + "learning_rate": 3.9591315453384425e-06, + "loss": 0.0235, + "step": 1678 + }, + { + "epoch": 28.96551724137931, + "grad_norm": 0.016247062012553215, + "learning_rate": 3.8314176245210725e-06, + "loss": 0.0005, + "step": 1680 + }, + { + "epoch": 29.0, + "grad_norm": 0.02568941004574299, + "learning_rate": 3.7037037037037037e-06, + "loss": 0.0007, + "step": 1682 + }, + { + "epoch": 29.0, + "eval_accuracy": 0.6742424242424242, + "eval_f1_macro": 0.5835078400423941, + "eval_f1_micro": 0.6742424242424242, + "eval_f1_weighted": 0.6678458659891735, + "eval_loss": 1.5345948934555054, + "eval_precision_macro": 0.5786037923941507, + "eval_precision_micro": 0.6742424242424242, + "eval_precision_weighted": 0.6702666872171518, + "eval_recall_macro": 0.5965382205513785, + "eval_recall_micro": 0.6742424242424242, + "eval_recall_weighted": 0.6742424242424242, + "eval_runtime": 2.0037, + "eval_samples_per_second": 65.879, + "eval_steps_per_second": 8.484, + "step": 1682 + }, + { + "epoch": 29.03448275862069, + "grad_norm": 0.007548394612967968, + "learning_rate": 3.5759897828863346e-06, + "loss": 0.0008, + "step": 1684 + }, + { + "epoch": 29.06896551724138, + "grad_norm": 0.11199589818716049, + "learning_rate": 3.448275862068966e-06, + "loss": 0.0011, + "step": 1686 + }, + { + "epoch": 29.103448275862068, + "grad_norm": 0.01992984302341938, + "learning_rate": 3.3205619412515967e-06, + "loss": 0.0005, + "step": 1688 + }, + { + "epoch": 29.137931034482758, + "grad_norm": 0.012301865965127945, + "learning_rate": 3.1928480204342275e-06, + "loss": 0.0006, + "step": 1690 + }, + { + "epoch": 29.17241379310345, + "grad_norm": 0.02116016112267971, + "learning_rate": 3.0651340996168583e-06, + "loss": 0.0097, + "step": 1692 + }, + { + "epoch": 29.20689655172414, + "grad_norm": 0.020142151042819023, + "learning_rate": 2.937420178799489e-06, + "loss": 0.0006, + "step": 1694 + }, + { + "epoch": 29.24137931034483, + "grad_norm": 0.042330626398324966, + "learning_rate": 2.80970625798212e-06, + "loss": 0.0009, + "step": 1696 + }, + { + "epoch": 29.275862068965516, + "grad_norm": 2.6495234966278076, + "learning_rate": 2.6819923371647512e-06, + "loss": 0.0576, + "step": 1698 + }, + { + "epoch": 29.310344827586206, + "grad_norm": 0.06072988361120224, + "learning_rate": 2.554278416347382e-06, + "loss": 0.0014, + "step": 1700 + }, + { + "epoch": 29.344827586206897, + "grad_norm": 0.19377738237380981, + "learning_rate": 2.426564495530013e-06, + "loss": 0.0018, + "step": 1702 + }, + { + "epoch": 29.379310344827587, + "grad_norm": 0.05790168046951294, + "learning_rate": 2.2988505747126437e-06, + "loss": 0.0007, + "step": 1704 + }, + { + "epoch": 29.413793103448278, + "grad_norm": 0.03023909032344818, + "learning_rate": 2.1711366538952746e-06, + "loss": 0.04, + "step": 1706 + }, + { + "epoch": 29.448275862068964, + "grad_norm": 0.039096005260944366, + "learning_rate": 2.043422733077906e-06, + "loss": 0.0009, + "step": 1708 + }, + { + "epoch": 29.482758620689655, + "grad_norm": 1.9999027252197266, + "learning_rate": 1.9157088122605362e-06, + "loss": 0.0383, + "step": 1710 + }, + { + "epoch": 29.517241379310345, + "grad_norm": 0.01942484639585018, + "learning_rate": 1.7879948914431673e-06, + "loss": 0.0004, + "step": 1712 + }, + { + "epoch": 29.551724137931036, + "grad_norm": 0.019709143787622452, + "learning_rate": 1.6602809706257983e-06, + "loss": 0.0006, + "step": 1714 + }, + { + "epoch": 29.586206896551722, + "grad_norm": 0.01487061008810997, + "learning_rate": 1.5325670498084292e-06, + "loss": 0.0003, + "step": 1716 + }, + { + "epoch": 29.620689655172413, + "grad_norm": 0.0076087117195129395, + "learning_rate": 1.40485312899106e-06, + "loss": 0.0003, + "step": 1718 + }, + { + "epoch": 29.655172413793103, + "grad_norm": 0.01488451100885868, + "learning_rate": 1.277139208173691e-06, + "loss": 0.0005, + "step": 1720 + }, + { + "epoch": 29.689655172413794, + "grad_norm": 0.00445834593847394, + "learning_rate": 1.1494252873563219e-06, + "loss": 0.0004, + "step": 1722 + }, + { + "epoch": 29.724137931034484, + "grad_norm": 0.01872408203780651, + "learning_rate": 1.021711366538953e-06, + "loss": 0.0005, + "step": 1724 + }, + { + "epoch": 29.75862068965517, + "grad_norm": 0.010615495964884758, + "learning_rate": 8.939974457215836e-07, + "loss": 0.0003, + "step": 1726 + }, + { + "epoch": 29.79310344827586, + "grad_norm": 0.015271478332579136, + "learning_rate": 7.662835249042146e-07, + "loss": 0.0005, + "step": 1728 + }, + { + "epoch": 29.82758620689655, + "grad_norm": 0.2127443104982376, + "learning_rate": 6.385696040868455e-07, + "loss": 0.002, + "step": 1730 + }, + { + "epoch": 29.862068965517242, + "grad_norm": 0.01297774724662304, + "learning_rate": 5.108556832694765e-07, + "loss": 0.0004, + "step": 1732 + }, + { + "epoch": 29.896551724137932, + "grad_norm": 0.010099313221871853, + "learning_rate": 3.831417624521073e-07, + "loss": 0.0064, + "step": 1734 + }, + { + "epoch": 29.93103448275862, + "grad_norm": 0.03221974894404411, + "learning_rate": 2.5542784163473823e-07, + "loss": 0.0005, + "step": 1736 + }, + { + "epoch": 29.96551724137931, + "grad_norm": 0.11465907841920853, + "learning_rate": 1.2771392081736911e-07, + "loss": 0.0009, + "step": 1738 + }, + { + "epoch": 30.0, + "grad_norm": 0.026466330513358116, + "learning_rate": 0.0, + "loss": 0.0004, + "step": 1740 + }, + { + "epoch": 30.0, + "eval_accuracy": 0.6742424242424242, + "eval_f1_macro": 0.5791488237085879, + "eval_f1_micro": 0.6742424242424242, + "eval_f1_weighted": 0.6661361118395798, + "eval_loss": 1.5231688022613525, + "eval_precision_macro": 0.5707257261879112, + "eval_precision_micro": 0.6742424242424242, + "eval_precision_weighted": 0.6628039716275012, + "eval_recall_macro": 0.5917763157894738, + "eval_recall_micro": 0.6742424242424242, + "eval_recall_weighted": 0.6742424242424242, + "eval_runtime": 2.1557, + "eval_samples_per_second": 61.234, + "eval_steps_per_second": 7.886, + "step": 1740 + }, + { + "epoch": 30.0, + "step": 1740, + "total_flos": 1.0740871074163507e+18, + "train_loss": 0.5156426236089975, + "train_runtime": 378.1612, + "train_samples_per_second": 36.651, + "train_steps_per_second": 4.601 + } + ], + "logging_steps": 2, + "max_steps": 1740, + "num_input_tokens_seen": 0, + "num_train_epochs": 30, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.0740871074163507e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}