{ "best_metric": 1.3132938146591187, "best_model_checkpoint": "WinKawaks/vit-tiny-patch16-224/checkpoint-319", "epoch": 25.0, "eval_steps": 500, "global_step": 725, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.06896551724137931, "grad_norm": 50.05537033081055, "learning_rate": 2.7397260273972604e-06, "loss": 2.392, "step": 2 }, { "epoch": 0.13793103448275862, "grad_norm": 21.303316116333008, "learning_rate": 5.479452054794521e-06, "loss": 1.993, "step": 4 }, { "epoch": 0.20689655172413793, "grad_norm": 29.138675689697266, "learning_rate": 8.21917808219178e-06, "loss": 1.9911, "step": 6 }, { "epoch": 0.27586206896551724, "grad_norm": 24.99004364013672, "learning_rate": 1.0958904109589042e-05, "loss": 2.1828, "step": 8 }, { "epoch": 0.3448275862068966, "grad_norm": 29.86865234375, "learning_rate": 1.3698630136986302e-05, "loss": 1.8818, "step": 10 }, { "epoch": 0.41379310344827586, "grad_norm": 33.12383270263672, "learning_rate": 1.643835616438356e-05, "loss": 2.0859, "step": 12 }, { "epoch": 0.4827586206896552, "grad_norm": 25.435964584350586, "learning_rate": 1.9178082191780822e-05, "loss": 2.0565, "step": 14 }, { "epoch": 0.5517241379310345, "grad_norm": 21.697423934936523, "learning_rate": 2.1917808219178083e-05, "loss": 1.9561, "step": 16 }, { "epoch": 0.6206896551724138, "grad_norm": 21.292062759399414, "learning_rate": 2.4657534246575342e-05, "loss": 1.8509, "step": 18 }, { "epoch": 0.6896551724137931, "grad_norm": 25.63444709777832, "learning_rate": 2.7397260273972603e-05, "loss": 1.9259, "step": 20 }, { "epoch": 0.7586206896551724, "grad_norm": 12.96942138671875, "learning_rate": 3.0136986301369862e-05, "loss": 1.8952, "step": 22 }, { "epoch": 0.8275862068965517, "grad_norm": 13.86652660369873, "learning_rate": 3.287671232876712e-05, "loss": 1.9227, "step": 24 }, { "epoch": 0.896551724137931, "grad_norm": 28.36823081970215, "learning_rate": 3.561643835616438e-05, "loss": 1.8958, "step": 26 }, { "epoch": 0.9655172413793104, "grad_norm": 18.257474899291992, "learning_rate": 3.8356164383561644e-05, "loss": 1.9719, "step": 28 }, { "epoch": 1.0, "eval_accuracy": 0.2196969696969697, "eval_f1_macro": 0.09493363779078065, "eval_f1_micro": 0.2196969696969697, "eval_f1_weighted": 0.11546937910574274, "eval_loss": 1.9209331274032593, "eval_precision_macro": 0.08906952611553096, "eval_precision_micro": 0.2196969696969697, "eval_precision_weighted": 0.10509758602978943, "eval_recall_macro": 0.17224489795918368, "eval_recall_micro": 0.2196969696969697, "eval_recall_weighted": 0.2196969696969697, "eval_runtime": 2.2105, "eval_samples_per_second": 59.716, "eval_steps_per_second": 7.691, "step": 29 }, { "epoch": 1.0344827586206897, "grad_norm": 28.444780349731445, "learning_rate": 4.1095890410958905e-05, "loss": 1.8292, "step": 30 }, { "epoch": 1.103448275862069, "grad_norm": 16.733516693115234, "learning_rate": 4.383561643835617e-05, "loss": 1.8189, "step": 32 }, { "epoch": 1.1724137931034484, "grad_norm": 16.255664825439453, "learning_rate": 4.657534246575342e-05, "loss": 1.8846, "step": 34 }, { "epoch": 1.2413793103448276, "grad_norm": 15.87771987915039, "learning_rate": 4.9315068493150684e-05, "loss": 1.9816, "step": 36 }, { "epoch": 1.3103448275862069, "grad_norm": 20.748777389526367, "learning_rate": 5.2054794520547945e-05, "loss": 1.8633, "step": 38 }, { "epoch": 1.3793103448275863, "grad_norm": 23.478870391845703, "learning_rate": 5.479452054794521e-05, "loss": 1.9196, "step": 40 }, { "epoch": 1.4482758620689655, "grad_norm": 14.190969467163086, "learning_rate": 5.753424657534247e-05, "loss": 1.8152, "step": 42 }, { "epoch": 1.5172413793103448, "grad_norm": 27.225452423095703, "learning_rate": 6.0273972602739724e-05, "loss": 1.879, "step": 44 }, { "epoch": 1.5862068965517242, "grad_norm": 16.791467666625977, "learning_rate": 6.301369863013699e-05, "loss": 1.851, "step": 46 }, { "epoch": 1.6551724137931034, "grad_norm": 17.566041946411133, "learning_rate": 6.575342465753424e-05, "loss": 1.8164, "step": 48 }, { "epoch": 1.7241379310344827, "grad_norm": 24.76405906677246, "learning_rate": 6.84931506849315e-05, "loss": 2.1271, "step": 50 }, { "epoch": 1.793103448275862, "grad_norm": 21.847782135009766, "learning_rate": 7.123287671232876e-05, "loss": 1.7536, "step": 52 }, { "epoch": 1.8620689655172413, "grad_norm": 19.893037796020508, "learning_rate": 7.397260273972603e-05, "loss": 1.7816, "step": 54 }, { "epoch": 1.9310344827586206, "grad_norm": 14.866455078125, "learning_rate": 7.671232876712329e-05, "loss": 1.5881, "step": 56 }, { "epoch": 2.0, "grad_norm": 20.373685836791992, "learning_rate": 7.945205479452055e-05, "loss": 1.8717, "step": 58 }, { "epoch": 2.0, "eval_accuracy": 0.19696969696969696, "eval_f1_macro": 0.09534116676973821, "eval_f1_micro": 0.19696969696969696, "eval_f1_weighted": 0.10689689098780009, "eval_loss": 2.0377984046936035, "eval_precision_macro": 0.19963369963369962, "eval_precision_micro": 0.19696969696969696, "eval_precision_weighted": 0.266025641025641, "eval_recall_macro": 0.17944066515495088, "eval_recall_micro": 0.19696969696969696, "eval_recall_weighted": 0.19696969696969696, "eval_runtime": 2.1913, "eval_samples_per_second": 60.238, "eval_steps_per_second": 7.758, "step": 58 }, { "epoch": 2.0689655172413794, "grad_norm": 21.545093536376953, "learning_rate": 8.219178082191781e-05, "loss": 1.752, "step": 60 }, { "epoch": 2.1379310344827585, "grad_norm": 15.046628952026367, "learning_rate": 8.493150684931507e-05, "loss": 1.7137, "step": 62 }, { "epoch": 2.206896551724138, "grad_norm": 21.083383560180664, "learning_rate": 8.767123287671233e-05, "loss": 1.7003, "step": 64 }, { "epoch": 2.2758620689655173, "grad_norm": 37.4631462097168, "learning_rate": 9.041095890410958e-05, "loss": 1.9736, "step": 66 }, { "epoch": 2.344827586206897, "grad_norm": 19.171171188354492, "learning_rate": 9.315068493150684e-05, "loss": 1.7332, "step": 68 }, { "epoch": 2.413793103448276, "grad_norm": 16.743974685668945, "learning_rate": 9.58904109589041e-05, "loss": 1.819, "step": 70 }, { "epoch": 2.4827586206896552, "grad_norm": 20.756649017333984, "learning_rate": 9.863013698630137e-05, "loss": 1.9201, "step": 72 }, { "epoch": 2.5517241379310347, "grad_norm": 17.703954696655273, "learning_rate": 9.984662576687117e-05, "loss": 1.6147, "step": 74 }, { "epoch": 2.6206896551724137, "grad_norm": 32.903743743896484, "learning_rate": 9.95398773006135e-05, "loss": 2.0663, "step": 76 }, { "epoch": 2.689655172413793, "grad_norm": 23.617671966552734, "learning_rate": 9.923312883435584e-05, "loss": 1.9193, "step": 78 }, { "epoch": 2.7586206896551726, "grad_norm": 30.0602970123291, "learning_rate": 9.892638036809816e-05, "loss": 1.7659, "step": 80 }, { "epoch": 2.8275862068965516, "grad_norm": 22.806859970092773, "learning_rate": 9.861963190184049e-05, "loss": 1.4494, "step": 82 }, { "epoch": 2.896551724137931, "grad_norm": 16.40064811706543, "learning_rate": 9.831288343558283e-05, "loss": 1.6664, "step": 84 }, { "epoch": 2.9655172413793105, "grad_norm": 24.703716278076172, "learning_rate": 9.800613496932515e-05, "loss": 1.9326, "step": 86 }, { "epoch": 3.0, "eval_accuracy": 0.3939393939393939, "eval_f1_macro": 0.2290203682353352, "eval_f1_micro": 0.3939393939393939, "eval_f1_weighted": 0.2938602829468567, "eval_loss": 1.7679647207260132, "eval_precision_macro": 0.21510735641170425, "eval_precision_micro": 0.3939393939393939, "eval_precision_weighted": 0.268160089800406, "eval_recall_macro": 0.30038548752834465, "eval_recall_micro": 0.3939393939393939, "eval_recall_weighted": 0.3939393939393939, "eval_runtime": 2.411, "eval_samples_per_second": 54.749, "eval_steps_per_second": 7.051, "step": 87 }, { "epoch": 3.0344827586206895, "grad_norm": 34.48782730102539, "learning_rate": 9.76993865030675e-05, "loss": 1.9289, "step": 88 }, { "epoch": 3.103448275862069, "grad_norm": 20.898527145385742, "learning_rate": 9.739263803680982e-05, "loss": 1.5641, "step": 90 }, { "epoch": 3.1724137931034484, "grad_norm": 20.089338302612305, "learning_rate": 9.708588957055215e-05, "loss": 1.6451, "step": 92 }, { "epoch": 3.2413793103448274, "grad_norm": 21.63447380065918, "learning_rate": 9.677914110429448e-05, "loss": 1.5942, "step": 94 }, { "epoch": 3.310344827586207, "grad_norm": 18.611299514770508, "learning_rate": 9.647239263803681e-05, "loss": 1.4619, "step": 96 }, { "epoch": 3.3793103448275863, "grad_norm": 14.333738327026367, "learning_rate": 9.616564417177915e-05, "loss": 1.6001, "step": 98 }, { "epoch": 3.4482758620689653, "grad_norm": 16.133724212646484, "learning_rate": 9.585889570552147e-05, "loss": 1.3741, "step": 100 }, { "epoch": 3.5172413793103448, "grad_norm": 29.28097915649414, "learning_rate": 9.555214723926381e-05, "loss": 1.6927, "step": 102 }, { "epoch": 3.586206896551724, "grad_norm": 17.335416793823242, "learning_rate": 9.524539877300614e-05, "loss": 1.1759, "step": 104 }, { "epoch": 3.655172413793103, "grad_norm": 25.239665985107422, "learning_rate": 9.493865030674846e-05, "loss": 1.4696, "step": 106 }, { "epoch": 3.7241379310344827, "grad_norm": 19.4155330657959, "learning_rate": 9.46319018404908e-05, "loss": 1.0003, "step": 108 }, { "epoch": 3.793103448275862, "grad_norm": 20.292905807495117, "learning_rate": 9.432515337423313e-05, "loss": 1.3851, "step": 110 }, { "epoch": 3.862068965517241, "grad_norm": 31.247602462768555, "learning_rate": 9.401840490797547e-05, "loss": 1.6519, "step": 112 }, { "epoch": 3.9310344827586206, "grad_norm": 25.56925392150879, "learning_rate": 9.37116564417178e-05, "loss": 1.4111, "step": 114 }, { "epoch": 4.0, "grad_norm": 23.311908721923828, "learning_rate": 9.340490797546013e-05, "loss": 1.2873, "step": 116 }, { "epoch": 4.0, "eval_accuracy": 0.44696969696969696, "eval_f1_macro": 0.35023855102631046, "eval_f1_micro": 0.44696969696969696, "eval_f1_weighted": 0.40824342918240536, "eval_loss": 1.5892395973205566, "eval_precision_macro": 0.4831240188383045, "eval_precision_micro": 0.44696969696969696, "eval_precision_weighted": 0.513986013986014, "eval_recall_macro": 0.36461829176114885, "eval_recall_micro": 0.44696969696969696, "eval_recall_weighted": 0.44696969696969696, "eval_runtime": 2.1869, "eval_samples_per_second": 60.359, "eval_steps_per_second": 7.774, "step": 116 }, { "epoch": 4.068965517241379, "grad_norm": 31.80891227722168, "learning_rate": 9.309815950920246e-05, "loss": 1.2896, "step": 118 }, { "epoch": 4.137931034482759, "grad_norm": 22.716781616210938, "learning_rate": 9.279141104294478e-05, "loss": 1.2486, "step": 120 }, { "epoch": 4.206896551724138, "grad_norm": 23.80237579345703, "learning_rate": 9.248466257668712e-05, "loss": 1.3835, "step": 122 }, { "epoch": 4.275862068965517, "grad_norm": 18.24744987487793, "learning_rate": 9.217791411042945e-05, "loss": 0.9575, "step": 124 }, { "epoch": 4.344827586206897, "grad_norm": 22.790117263793945, "learning_rate": 9.187116564417179e-05, "loss": 1.3711, "step": 126 }, { "epoch": 4.413793103448276, "grad_norm": 15.804698944091797, "learning_rate": 9.156441717791411e-05, "loss": 1.4226, "step": 128 }, { "epoch": 4.482758620689655, "grad_norm": 30.167261123657227, "learning_rate": 9.125766871165644e-05, "loss": 1.7736, "step": 130 }, { "epoch": 4.551724137931035, "grad_norm": 15.975086212158203, "learning_rate": 9.095092024539878e-05, "loss": 1.4395, "step": 132 }, { "epoch": 4.620689655172414, "grad_norm": 23.406415939331055, "learning_rate": 9.06441717791411e-05, "loss": 1.6491, "step": 134 }, { "epoch": 4.689655172413794, "grad_norm": 30.07583236694336, "learning_rate": 9.033742331288344e-05, "loss": 1.747, "step": 136 }, { "epoch": 4.758620689655173, "grad_norm": 20.77846336364746, "learning_rate": 9.003067484662577e-05, "loss": 1.6648, "step": 138 }, { "epoch": 4.827586206896552, "grad_norm": 11.187516212463379, "learning_rate": 8.972392638036811e-05, "loss": 1.1932, "step": 140 }, { "epoch": 4.896551724137931, "grad_norm": 15.845014572143555, "learning_rate": 8.941717791411043e-05, "loss": 1.4271, "step": 142 }, { "epoch": 4.9655172413793105, "grad_norm": 15.977095603942871, "learning_rate": 8.911042944785276e-05, "loss": 1.3997, "step": 144 }, { "epoch": 5.0, "eval_accuracy": 0.5, "eval_f1_macro": 0.34814087704047136, "eval_f1_micro": 0.5, "eval_f1_weighted": 0.4244756131292643, "eval_loss": 1.4773013591766357, "eval_precision_macro": 0.34626430480089015, "eval_precision_micro": 0.5, "eval_precision_weighted": 0.41194737757930666, "eval_recall_macro": 0.4051700680272109, "eval_recall_micro": 0.5, "eval_recall_weighted": 0.5, "eval_runtime": 2.2186, "eval_samples_per_second": 59.497, "eval_steps_per_second": 7.662, "step": 145 }, { "epoch": 5.0344827586206895, "grad_norm": 18.65735626220703, "learning_rate": 8.88036809815951e-05, "loss": 1.6883, "step": 146 }, { "epoch": 5.103448275862069, "grad_norm": 18.897695541381836, "learning_rate": 8.849693251533742e-05, "loss": 1.3035, "step": 148 }, { "epoch": 5.172413793103448, "grad_norm": 20.19015884399414, "learning_rate": 8.819018404907976e-05, "loss": 1.1881, "step": 150 }, { "epoch": 5.241379310344827, "grad_norm": 18.574827194213867, "learning_rate": 8.788343558282209e-05, "loss": 1.0471, "step": 152 }, { "epoch": 5.310344827586207, "grad_norm": 15.6314115524292, "learning_rate": 8.757668711656443e-05, "loss": 1.2302, "step": 154 }, { "epoch": 5.379310344827586, "grad_norm": 20.703832626342773, "learning_rate": 8.726993865030675e-05, "loss": 1.1877, "step": 156 }, { "epoch": 5.448275862068965, "grad_norm": 28.778301239013672, "learning_rate": 8.696319018404908e-05, "loss": 1.5936, "step": 158 }, { "epoch": 5.517241379310345, "grad_norm": 25.026071548461914, "learning_rate": 8.665644171779142e-05, "loss": 1.0255, "step": 160 }, { "epoch": 5.586206896551724, "grad_norm": 17.47880744934082, "learning_rate": 8.634969325153374e-05, "loss": 1.3967, "step": 162 }, { "epoch": 5.655172413793103, "grad_norm": 20.092926025390625, "learning_rate": 8.604294478527608e-05, "loss": 1.2006, "step": 164 }, { "epoch": 5.724137931034483, "grad_norm": 17.11634635925293, "learning_rate": 8.573619631901841e-05, "loss": 1.1081, "step": 166 }, { "epoch": 5.793103448275862, "grad_norm": 21.26296043395996, "learning_rate": 8.542944785276073e-05, "loss": 1.0949, "step": 168 }, { "epoch": 5.862068965517241, "grad_norm": 31.731033325195312, "learning_rate": 8.512269938650307e-05, "loss": 1.3502, "step": 170 }, { "epoch": 5.931034482758621, "grad_norm": 25.80246353149414, "learning_rate": 8.48159509202454e-05, "loss": 1.5673, "step": 172 }, { "epoch": 6.0, "grad_norm": 23.60406494140625, "learning_rate": 8.450920245398774e-05, "loss": 1.7041, "step": 174 }, { "epoch": 6.0, "eval_accuracy": 0.5378787878787878, "eval_f1_macro": 0.4265559579914266, "eval_f1_micro": 0.5378787878787878, "eval_f1_weighted": 0.5005041399030636, "eval_loss": 1.4405734539031982, "eval_precision_macro": 0.5010893868036724, "eval_precision_micro": 0.5378787878787878, "eval_precision_weighted": 0.5628245491881856, "eval_recall_macro": 0.4529024943310657, "eval_recall_micro": 0.5378787878787878, "eval_recall_weighted": 0.5378787878787878, "eval_runtime": 2.215, "eval_samples_per_second": 59.594, "eval_steps_per_second": 7.675, "step": 174 }, { "epoch": 6.068965517241379, "grad_norm": 20.430105209350586, "learning_rate": 8.420245398773006e-05, "loss": 1.0018, "step": 176 }, { "epoch": 6.137931034482759, "grad_norm": 14.565896987915039, "learning_rate": 8.38957055214724e-05, "loss": 1.02, "step": 178 }, { "epoch": 6.206896551724138, "grad_norm": 21.370939254760742, "learning_rate": 8.358895705521473e-05, "loss": 1.1403, "step": 180 }, { "epoch": 6.275862068965517, "grad_norm": 16.42601776123047, "learning_rate": 8.328220858895705e-05, "loss": 1.4774, "step": 182 }, { "epoch": 6.344827586206897, "grad_norm": 18.844532012939453, "learning_rate": 8.297546012269939e-05, "loss": 1.1568, "step": 184 }, { "epoch": 6.413793103448276, "grad_norm": 12.009956359863281, "learning_rate": 8.266871165644172e-05, "loss": 0.8588, "step": 186 }, { "epoch": 6.482758620689655, "grad_norm": 21.886213302612305, "learning_rate": 8.236196319018406e-05, "loss": 0.8875, "step": 188 }, { "epoch": 6.551724137931035, "grad_norm": 16.68918800354004, "learning_rate": 8.205521472392638e-05, "loss": 1.2283, "step": 190 }, { "epoch": 6.620689655172414, "grad_norm": 19.978803634643555, "learning_rate": 8.174846625766872e-05, "loss": 1.0841, "step": 192 }, { "epoch": 6.689655172413794, "grad_norm": 16.348190307617188, "learning_rate": 8.144171779141105e-05, "loss": 0.9849, "step": 194 }, { "epoch": 6.758620689655173, "grad_norm": 21.0911865234375, "learning_rate": 8.113496932515337e-05, "loss": 1.6812, "step": 196 }, { "epoch": 6.827586206896552, "grad_norm": 17.614559173583984, "learning_rate": 8.082822085889571e-05, "loss": 1.0849, "step": 198 }, { "epoch": 6.896551724137931, "grad_norm": 19.585248947143555, "learning_rate": 8.052147239263804e-05, "loss": 0.9886, "step": 200 }, { "epoch": 6.9655172413793105, "grad_norm": 19.765750885009766, "learning_rate": 8.021472392638038e-05, "loss": 1.1863, "step": 202 }, { "epoch": 7.0, "eval_accuracy": 0.5681818181818182, "eval_f1_macro": 0.4758501564659626, "eval_f1_micro": 0.5681818181818182, "eval_f1_weighted": 0.5400274091359906, "eval_loss": 1.3679978847503662, "eval_precision_macro": 0.5558842701699845, "eval_precision_micro": 0.5681818181818182, "eval_precision_weighted": 0.603162194071285, "eval_recall_macro": 0.4831443688586546, "eval_recall_micro": 0.5681818181818182, "eval_recall_weighted": 0.5681818181818182, "eval_runtime": 2.1971, "eval_samples_per_second": 60.078, "eval_steps_per_second": 7.737, "step": 203 }, { "epoch": 7.0344827586206895, "grad_norm": 12.333309173583984, "learning_rate": 7.99079754601227e-05, "loss": 0.8463, "step": 204 }, { "epoch": 7.103448275862069, "grad_norm": 21.288188934326172, "learning_rate": 7.960122699386503e-05, "loss": 0.9617, "step": 206 }, { "epoch": 7.172413793103448, "grad_norm": 23.408567428588867, "learning_rate": 7.929447852760737e-05, "loss": 1.2719, "step": 208 }, { "epoch": 7.241379310344827, "grad_norm": 16.838363647460938, "learning_rate": 7.898773006134969e-05, "loss": 0.9394, "step": 210 }, { "epoch": 7.310344827586207, "grad_norm": 19.406648635864258, "learning_rate": 7.868098159509203e-05, "loss": 1.0913, "step": 212 }, { "epoch": 7.379310344827586, "grad_norm": 25.161184310913086, "learning_rate": 7.837423312883436e-05, "loss": 1.029, "step": 214 }, { "epoch": 7.448275862068965, "grad_norm": 24.920177459716797, "learning_rate": 7.80674846625767e-05, "loss": 1.3211, "step": 216 }, { "epoch": 7.517241379310345, "grad_norm": 22.075044631958008, "learning_rate": 7.776073619631902e-05, "loss": 1.226, "step": 218 }, { "epoch": 7.586206896551724, "grad_norm": 17.07358169555664, "learning_rate": 7.745398773006135e-05, "loss": 0.8096, "step": 220 }, { "epoch": 7.655172413793103, "grad_norm": 23.150299072265625, "learning_rate": 7.714723926380369e-05, "loss": 1.0163, "step": 222 }, { "epoch": 7.724137931034483, "grad_norm": 19.737802505493164, "learning_rate": 7.684049079754601e-05, "loss": 1.0773, "step": 224 }, { "epoch": 7.793103448275862, "grad_norm": 25.407928466796875, "learning_rate": 7.653374233128835e-05, "loss": 1.2907, "step": 226 }, { "epoch": 7.862068965517241, "grad_norm": 14.86108112335205, "learning_rate": 7.622699386503068e-05, "loss": 0.9851, "step": 228 }, { "epoch": 7.931034482758621, "grad_norm": 16.41703987121582, "learning_rate": 7.5920245398773e-05, "loss": 1.405, "step": 230 }, { "epoch": 8.0, "grad_norm": 20.147233963012695, "learning_rate": 7.561349693251534e-05, "loss": 0.9817, "step": 232 }, { "epoch": 8.0, "eval_accuracy": 0.5227272727272727, "eval_f1_macro": 0.43986070618723677, "eval_f1_micro": 0.5227272727272727, "eval_f1_weighted": 0.4968812272383701, "eval_loss": 1.3514596223831177, "eval_precision_macro": 0.44448009061813915, "eval_precision_micro": 0.5227272727272727, "eval_precision_weighted": 0.5088011746058729, "eval_recall_macro": 0.47222222222222215, "eval_recall_micro": 0.5227272727272727, "eval_recall_weighted": 0.5227272727272727, "eval_runtime": 2.1836, "eval_samples_per_second": 60.451, "eval_steps_per_second": 7.785, "step": 232 }, { "epoch": 8.068965517241379, "grad_norm": 19.058284759521484, "learning_rate": 7.530674846625767e-05, "loss": 0.7432, "step": 234 }, { "epoch": 8.137931034482758, "grad_norm": 15.031048774719238, "learning_rate": 7.500000000000001e-05, "loss": 0.9929, "step": 236 }, { "epoch": 8.206896551724139, "grad_norm": 22.36937713623047, "learning_rate": 7.469325153374233e-05, "loss": 1.121, "step": 238 }, { "epoch": 8.275862068965518, "grad_norm": 20.049163818359375, "learning_rate": 7.438650306748467e-05, "loss": 1.0773, "step": 240 }, { "epoch": 8.344827586206897, "grad_norm": 17.548959732055664, "learning_rate": 7.4079754601227e-05, "loss": 0.6477, "step": 242 }, { "epoch": 8.413793103448276, "grad_norm": 25.496204376220703, "learning_rate": 7.377300613496932e-05, "loss": 0.8577, "step": 244 }, { "epoch": 8.482758620689655, "grad_norm": 22.851713180541992, "learning_rate": 7.346625766871166e-05, "loss": 1.0034, "step": 246 }, { "epoch": 8.551724137931034, "grad_norm": 26.218107223510742, "learning_rate": 7.315950920245399e-05, "loss": 0.7915, "step": 248 }, { "epoch": 8.620689655172415, "grad_norm": 18.867645263671875, "learning_rate": 7.285276073619633e-05, "loss": 1.0731, "step": 250 }, { "epoch": 8.689655172413794, "grad_norm": 16.624637603759766, "learning_rate": 7.254601226993865e-05, "loss": 0.706, "step": 252 }, { "epoch": 8.758620689655173, "grad_norm": 26.590402603149414, "learning_rate": 7.223926380368099e-05, "loss": 1.1713, "step": 254 }, { "epoch": 8.827586206896552, "grad_norm": 26.262710571289062, "learning_rate": 7.193251533742332e-05, "loss": 0.8558, "step": 256 }, { "epoch": 8.89655172413793, "grad_norm": 24.299407958984375, "learning_rate": 7.162576687116564e-05, "loss": 1.2063, "step": 258 }, { "epoch": 8.96551724137931, "grad_norm": 14.688630104064941, "learning_rate": 7.131901840490798e-05, "loss": 0.617, "step": 260 }, { "epoch": 9.0, "eval_accuracy": 0.5909090909090909, "eval_f1_macro": 0.48949553001277135, "eval_f1_micro": 0.5909090909090909, "eval_f1_weighted": 0.5554842002399473, "eval_loss": 1.3866709470748901, "eval_precision_macro": 0.5135558290637433, "eval_precision_micro": 0.5909090909090909, "eval_precision_weighted": 0.5775828309138267, "eval_recall_macro": 0.5183068783068784, "eval_recall_micro": 0.5909090909090909, "eval_recall_weighted": 0.5909090909090909, "eval_runtime": 2.1836, "eval_samples_per_second": 60.452, "eval_steps_per_second": 7.785, "step": 261 }, { "epoch": 9.03448275862069, "grad_norm": 14.76333236694336, "learning_rate": 7.101226993865031e-05, "loss": 0.5944, "step": 262 }, { "epoch": 9.10344827586207, "grad_norm": 24.155582427978516, "learning_rate": 7.070552147239265e-05, "loss": 0.7582, "step": 264 }, { "epoch": 9.172413793103448, "grad_norm": 28.48207664489746, "learning_rate": 7.039877300613497e-05, "loss": 0.8912, "step": 266 }, { "epoch": 9.241379310344827, "grad_norm": 10.216355323791504, "learning_rate": 7.00920245398773e-05, "loss": 0.6462, "step": 268 }, { "epoch": 9.310344827586206, "grad_norm": 26.633636474609375, "learning_rate": 6.978527607361964e-05, "loss": 0.9824, "step": 270 }, { "epoch": 9.379310344827585, "grad_norm": 24.09172821044922, "learning_rate": 6.947852760736196e-05, "loss": 0.6394, "step": 272 }, { "epoch": 9.448275862068966, "grad_norm": 27.006250381469727, "learning_rate": 6.91717791411043e-05, "loss": 0.9335, "step": 274 }, { "epoch": 9.517241379310345, "grad_norm": 22.846731185913086, "learning_rate": 6.886503067484663e-05, "loss": 0.6089, "step": 276 }, { "epoch": 9.586206896551724, "grad_norm": 19.58112144470215, "learning_rate": 6.855828220858897e-05, "loss": 0.7781, "step": 278 }, { "epoch": 9.655172413793103, "grad_norm": 13.92150592803955, "learning_rate": 6.825153374233129e-05, "loss": 0.5359, "step": 280 }, { "epoch": 9.724137931034482, "grad_norm": 31.10985565185547, "learning_rate": 6.794478527607362e-05, "loss": 1.1988, "step": 282 }, { "epoch": 9.793103448275861, "grad_norm": 23.35214614868164, "learning_rate": 6.763803680981596e-05, "loss": 0.9214, "step": 284 }, { "epoch": 9.862068965517242, "grad_norm": 28.431663513183594, "learning_rate": 6.733128834355828e-05, "loss": 0.7769, "step": 286 }, { "epoch": 9.931034482758621, "grad_norm": 26.589706420898438, "learning_rate": 6.702453987730062e-05, "loss": 0.7962, "step": 288 }, { "epoch": 10.0, "grad_norm": 31.327491760253906, "learning_rate": 6.671779141104295e-05, "loss": 1.0365, "step": 290 }, { "epoch": 10.0, "eval_accuracy": 0.5378787878787878, "eval_f1_macro": 0.43132712141928736, "eval_f1_micro": 0.5378787878787878, "eval_f1_weighted": 0.49605830250991545, "eval_loss": 1.4607229232788086, "eval_precision_macro": 0.4370555865025911, "eval_precision_micro": 0.5378787878787878, "eval_precision_weighted": 0.4996555594942692, "eval_recall_macro": 0.46741496598639454, "eval_recall_micro": 0.5378787878787878, "eval_recall_weighted": 0.5378787878787878, "eval_runtime": 2.1864, "eval_samples_per_second": 60.372, "eval_steps_per_second": 7.775, "step": 290 }, { "epoch": 10.068965517241379, "grad_norm": 35.857444763183594, "learning_rate": 6.641104294478529e-05, "loss": 1.1976, "step": 292 }, { "epoch": 10.137931034482758, "grad_norm": 19.138635635375977, "learning_rate": 6.610429447852761e-05, "loss": 0.775, "step": 294 }, { "epoch": 10.206896551724139, "grad_norm": 28.3044490814209, "learning_rate": 6.579754601226994e-05, "loss": 0.7642, "step": 296 }, { "epoch": 10.275862068965518, "grad_norm": 20.905742645263672, "learning_rate": 6.549079754601228e-05, "loss": 0.5362, "step": 298 }, { "epoch": 10.344827586206897, "grad_norm": 18.387508392333984, "learning_rate": 6.51840490797546e-05, "loss": 0.9838, "step": 300 }, { "epoch": 10.413793103448276, "grad_norm": 13.722414016723633, "learning_rate": 6.487730061349694e-05, "loss": 0.446, "step": 302 }, { "epoch": 10.482758620689655, "grad_norm": 20.892261505126953, "learning_rate": 6.457055214723927e-05, "loss": 0.7612, "step": 304 }, { "epoch": 10.551724137931034, "grad_norm": 13.154946327209473, "learning_rate": 6.426380368098159e-05, "loss": 0.7933, "step": 306 }, { "epoch": 10.620689655172415, "grad_norm": 16.057727813720703, "learning_rate": 6.395705521472393e-05, "loss": 0.522, "step": 308 }, { "epoch": 10.689655172413794, "grad_norm": 19.725608825683594, "learning_rate": 6.365030674846626e-05, "loss": 0.7614, "step": 310 }, { "epoch": 10.758620689655173, "grad_norm": 24.844079971313477, "learning_rate": 6.33435582822086e-05, "loss": 0.7933, "step": 312 }, { "epoch": 10.827586206896552, "grad_norm": 18.63338279724121, "learning_rate": 6.303680981595092e-05, "loss": 0.7279, "step": 314 }, { "epoch": 10.89655172413793, "grad_norm": 21.24047088623047, "learning_rate": 6.273006134969326e-05, "loss": 0.5442, "step": 316 }, { "epoch": 10.96551724137931, "grad_norm": 33.37268829345703, "learning_rate": 6.242331288343559e-05, "loss": 0.6815, "step": 318 }, { "epoch": 11.0, "eval_accuracy": 0.5909090909090909, "eval_f1_macro": 0.49620898933890345, "eval_f1_micro": 0.5909090909090909, "eval_f1_weighted": 0.5663560659976746, "eval_loss": 1.3132938146591187, "eval_precision_macro": 0.5087159863945577, "eval_precision_micro": 0.5909090909090909, "eval_precision_weighted": 0.5741680194805194, "eval_recall_macro": 0.5132879818594104, "eval_recall_micro": 0.5909090909090909, "eval_recall_weighted": 0.5909090909090909, "eval_runtime": 2.1795, "eval_samples_per_second": 60.565, "eval_steps_per_second": 7.8, "step": 319 }, { "epoch": 11.03448275862069, "grad_norm": 16.19150733947754, "learning_rate": 6.211656441717791e-05, "loss": 0.644, "step": 320 }, { "epoch": 11.10344827586207, "grad_norm": 20.564546585083008, "learning_rate": 6.180981595092025e-05, "loss": 0.4998, "step": 322 }, { "epoch": 11.172413793103448, "grad_norm": 19.2364444732666, "learning_rate": 6.150306748466258e-05, "loss": 0.7469, "step": 324 }, { "epoch": 11.241379310344827, "grad_norm": 11.9139404296875, "learning_rate": 6.119631901840492e-05, "loss": 0.3421, "step": 326 }, { "epoch": 11.310344827586206, "grad_norm": 15.564549446105957, "learning_rate": 6.088957055214725e-05, "loss": 0.454, "step": 328 }, { "epoch": 11.379310344827585, "grad_norm": 15.790903091430664, "learning_rate": 6.058282208588958e-05, "loss": 0.5501, "step": 330 }, { "epoch": 11.448275862068966, "grad_norm": 20.395984649658203, "learning_rate": 6.02760736196319e-05, "loss": 0.7396, "step": 332 }, { "epoch": 11.517241379310345, "grad_norm": 13.017558097839355, "learning_rate": 5.996932515337423e-05, "loss": 0.3902, "step": 334 }, { "epoch": 11.586206896551724, "grad_norm": 16.60504150390625, "learning_rate": 5.9662576687116564e-05, "loss": 0.4817, "step": 336 }, { "epoch": 11.655172413793103, "grad_norm": 26.98207664489746, "learning_rate": 5.93558282208589e-05, "loss": 0.7637, "step": 338 }, { "epoch": 11.724137931034482, "grad_norm": 27.36790657043457, "learning_rate": 5.9049079754601235e-05, "loss": 0.7137, "step": 340 }, { "epoch": 11.793103448275861, "grad_norm": 21.537046432495117, "learning_rate": 5.874233128834357e-05, "loss": 0.5965, "step": 342 }, { "epoch": 11.862068965517242, "grad_norm": 23.125181198120117, "learning_rate": 5.8435582822085886e-05, "loss": 0.5142, "step": 344 }, { "epoch": 11.931034482758621, "grad_norm": 24.718408584594727, "learning_rate": 5.812883435582822e-05, "loss": 0.7012, "step": 346 }, { "epoch": 12.0, "grad_norm": 10.676513671875, "learning_rate": 5.782208588957055e-05, "loss": 0.4153, "step": 348 }, { "epoch": 12.0, "eval_accuracy": 0.5909090909090909, "eval_f1_macro": 0.5082184346733783, "eval_f1_micro": 0.5909090909090909, "eval_f1_weighted": 0.5734953480846786, "eval_loss": 1.3527586460113525, "eval_precision_macro": 0.518512557765101, "eval_precision_micro": 0.5909090909090909, "eval_precision_weighted": 0.5819735429220373, "eval_recall_macro": 0.5201889644746788, "eval_recall_micro": 0.5909090909090909, "eval_recall_weighted": 0.5909090909090909, "eval_runtime": 2.1855, "eval_samples_per_second": 60.399, "eval_steps_per_second": 7.779, "step": 348 }, { "epoch": 12.068965517241379, "grad_norm": 9.67835521697998, "learning_rate": 5.751533742331289e-05, "loss": 0.341, "step": 350 }, { "epoch": 12.137931034482758, "grad_norm": 26.32645034790039, "learning_rate": 5.720858895705522e-05, "loss": 0.6657, "step": 352 }, { "epoch": 12.206896551724139, "grad_norm": 26.893024444580078, "learning_rate": 5.6901840490797555e-05, "loss": 0.6414, "step": 354 }, { "epoch": 12.275862068965518, "grad_norm": 20.325834274291992, "learning_rate": 5.6595092024539874e-05, "loss": 0.5375, "step": 356 }, { "epoch": 12.344827586206897, "grad_norm": 14.722708702087402, "learning_rate": 5.6288343558282206e-05, "loss": 0.3507, "step": 358 }, { "epoch": 12.413793103448276, "grad_norm": 16.164493560791016, "learning_rate": 5.598159509202454e-05, "loss": 0.3948, "step": 360 }, { "epoch": 12.482758620689655, "grad_norm": 20.786996841430664, "learning_rate": 5.567484662576688e-05, "loss": 0.4795, "step": 362 }, { "epoch": 12.551724137931034, "grad_norm": 14.129579544067383, "learning_rate": 5.536809815950921e-05, "loss": 0.2681, "step": 364 }, { "epoch": 12.620689655172415, "grad_norm": 14.56933879852295, "learning_rate": 5.506134969325154e-05, "loss": 0.5054, "step": 366 }, { "epoch": 12.689655172413794, "grad_norm": 30.525728225708008, "learning_rate": 5.475460122699386e-05, "loss": 0.5441, "step": 368 }, { "epoch": 12.758620689655173, "grad_norm": 26.576383590698242, "learning_rate": 5.4447852760736193e-05, "loss": 0.7958, "step": 370 }, { "epoch": 12.827586206896552, "grad_norm": 15.745062828063965, "learning_rate": 5.4141104294478526e-05, "loss": 0.4405, "step": 372 }, { "epoch": 12.89655172413793, "grad_norm": 27.86982536315918, "learning_rate": 5.3834355828220865e-05, "loss": 0.5312, "step": 374 }, { "epoch": 12.96551724137931, "grad_norm": 14.46611213684082, "learning_rate": 5.35276073619632e-05, "loss": 0.3396, "step": 376 }, { "epoch": 13.0, "eval_accuracy": 0.5909090909090909, "eval_f1_macro": 0.53715778106022, "eval_f1_micro": 0.5909090909090909, "eval_f1_weighted": 0.5830308957803414, "eval_loss": 1.385578989982605, "eval_precision_macro": 0.5622709610159853, "eval_precision_micro": 0.5909090909090909, "eval_precision_weighted": 0.6017770445498773, "eval_recall_macro": 0.5387226001511716, "eval_recall_micro": 0.5909090909090909, "eval_recall_weighted": 0.5909090909090909, "eval_runtime": 2.1898, "eval_samples_per_second": 60.279, "eval_steps_per_second": 7.763, "step": 377 }, { "epoch": 13.03448275862069, "grad_norm": 17.2612361907959, "learning_rate": 5.322085889570553e-05, "loss": 0.3936, "step": 378 }, { "epoch": 13.10344827586207, "grad_norm": 14.51452350616455, "learning_rate": 5.291411042944786e-05, "loss": 0.3816, "step": 380 }, { "epoch": 13.172413793103448, "grad_norm": 11.573400497436523, "learning_rate": 5.260736196319018e-05, "loss": 0.3908, "step": 382 }, { "epoch": 13.241379310344827, "grad_norm": 18.94368553161621, "learning_rate": 5.230061349693251e-05, "loss": 0.5848, "step": 384 }, { "epoch": 13.310344827586206, "grad_norm": 18.253276824951172, "learning_rate": 5.1993865030674845e-05, "loss": 0.2693, "step": 386 }, { "epoch": 13.379310344827585, "grad_norm": 12.632643699645996, "learning_rate": 5.1687116564417185e-05, "loss": 0.4368, "step": 388 }, { "epoch": 13.448275862068966, "grad_norm": 14.242535591125488, "learning_rate": 5.138036809815952e-05, "loss": 0.3045, "step": 390 }, { "epoch": 13.517241379310345, "grad_norm": 10.820467948913574, "learning_rate": 5.107361963190185e-05, "loss": 0.191, "step": 392 }, { "epoch": 13.586206896551724, "grad_norm": 18.95819091796875, "learning_rate": 5.076687116564417e-05, "loss": 0.5458, "step": 394 }, { "epoch": 13.655172413793103, "grad_norm": 21.91457748413086, "learning_rate": 5.04601226993865e-05, "loss": 0.7368, "step": 396 }, { "epoch": 13.724137931034482, "grad_norm": 28.396440505981445, "learning_rate": 5.015337423312883e-05, "loss": 0.6519, "step": 398 }, { "epoch": 13.793103448275861, "grad_norm": 16.6456356048584, "learning_rate": 4.984662576687117e-05, "loss": 0.5557, "step": 400 }, { "epoch": 13.862068965517242, "grad_norm": 19.1430606842041, "learning_rate": 4.9539877300613504e-05, "loss": 0.6529, "step": 402 }, { "epoch": 13.931034482758621, "grad_norm": 26.434980392456055, "learning_rate": 4.923312883435583e-05, "loss": 0.6911, "step": 404 }, { "epoch": 14.0, "grad_norm": 30.586454391479492, "learning_rate": 4.892638036809816e-05, "loss": 0.5415, "step": 406 }, { "epoch": 14.0, "eval_accuracy": 0.5909090909090909, "eval_f1_macro": 0.5132147277659137, "eval_f1_micro": 0.5909090909090909, "eval_f1_weighted": 0.5795267633305671, "eval_loss": 1.4251549243927002, "eval_precision_macro": 0.5222798718196696, "eval_precision_micro": 0.5909090909090909, "eval_precision_weighted": 0.5892947366820627, "eval_recall_macro": 0.5254724111866969, "eval_recall_micro": 0.5909090909090909, "eval_recall_weighted": 0.5909090909090909, "eval_runtime": 2.1927, "eval_samples_per_second": 60.199, "eval_steps_per_second": 7.753, "step": 406 }, { "epoch": 14.068965517241379, "grad_norm": 22.47860336303711, "learning_rate": 4.8619631901840495e-05, "loss": 0.2761, "step": 408 }, { "epoch": 14.137931034482758, "grad_norm": 6.592859745025635, "learning_rate": 4.831288343558282e-05, "loss": 0.2115, "step": 410 }, { "epoch": 14.206896551724139, "grad_norm": 11.651180267333984, "learning_rate": 4.800613496932516e-05, "loss": 0.2898, "step": 412 }, { "epoch": 14.275862068965518, "grad_norm": 24.14730453491211, "learning_rate": 4.769938650306749e-05, "loss": 0.468, "step": 414 }, { "epoch": 14.344827586206897, "grad_norm": 22.79234504699707, "learning_rate": 4.739263803680982e-05, "loss": 0.3194, "step": 416 }, { "epoch": 14.413793103448276, "grad_norm": 19.285917282104492, "learning_rate": 4.708588957055215e-05, "loss": 0.3259, "step": 418 }, { "epoch": 14.482758620689655, "grad_norm": 15.568497657775879, "learning_rate": 4.677914110429448e-05, "loss": 0.229, "step": 420 }, { "epoch": 14.551724137931034, "grad_norm": 6.664709091186523, "learning_rate": 4.647239263803681e-05, "loss": 0.5116, "step": 422 }, { "epoch": 14.620689655172415, "grad_norm": 27.09947967529297, "learning_rate": 4.616564417177914e-05, "loss": 0.3526, "step": 424 }, { "epoch": 14.689655172413794, "grad_norm": 19.279560089111328, "learning_rate": 4.585889570552148e-05, "loss": 0.327, "step": 426 }, { "epoch": 14.758620689655173, "grad_norm": 14.68875789642334, "learning_rate": 4.5552147239263805e-05, "loss": 0.1932, "step": 428 }, { "epoch": 14.827586206896552, "grad_norm": 10.188983917236328, "learning_rate": 4.524539877300614e-05, "loss": 0.3491, "step": 430 }, { "epoch": 14.89655172413793, "grad_norm": 30.82689094543457, "learning_rate": 4.493865030674847e-05, "loss": 0.5371, "step": 432 }, { "epoch": 14.96551724137931, "grad_norm": 25.854101181030273, "learning_rate": 4.4631901840490795e-05, "loss": 0.4421, "step": 434 }, { "epoch": 15.0, "eval_accuracy": 0.6136363636363636, "eval_f1_macro": 0.5574127938548423, "eval_f1_micro": 0.6136363636363636, "eval_f1_weighted": 0.6086044486243096, "eval_loss": 1.4080591201782227, "eval_precision_macro": 0.5752992105933282, "eval_precision_micro": 0.6136363636363636, "eval_precision_weighted": 0.6148821098687408, "eval_recall_macro": 0.5531670445956159, "eval_recall_micro": 0.6136363636363636, "eval_recall_weighted": 0.6136363636363636, "eval_runtime": 2.1892, "eval_samples_per_second": 60.296, "eval_steps_per_second": 7.765, "step": 435 }, { "epoch": 15.03448275862069, "grad_norm": 10.050946235656738, "learning_rate": 4.432515337423313e-05, "loss": 0.3007, "step": 436 }, { "epoch": 15.10344827586207, "grad_norm": 12.756734848022461, "learning_rate": 4.4018404907975466e-05, "loss": 0.3396, "step": 438 }, { "epoch": 15.172413793103448, "grad_norm": 25.455589294433594, "learning_rate": 4.371165644171779e-05, "loss": 0.3914, "step": 440 }, { "epoch": 15.241379310344827, "grad_norm": 13.843463897705078, "learning_rate": 4.3404907975460124e-05, "loss": 0.3308, "step": 442 }, { "epoch": 15.310344827586206, "grad_norm": 23.339752197265625, "learning_rate": 4.309815950920246e-05, "loss": 0.4584, "step": 444 }, { "epoch": 15.379310344827585, "grad_norm": 8.64341926574707, "learning_rate": 4.279141104294479e-05, "loss": 0.1732, "step": 446 }, { "epoch": 15.448275862068966, "grad_norm": 6.727046012878418, "learning_rate": 4.2484662576687115e-05, "loss": 0.2713, "step": 448 }, { "epoch": 15.517241379310345, "grad_norm": 30.494932174682617, "learning_rate": 4.2177914110429454e-05, "loss": 0.2759, "step": 450 }, { "epoch": 15.586206896551724, "grad_norm": 25.142616271972656, "learning_rate": 4.1871165644171786e-05, "loss": 0.2707, "step": 452 }, { "epoch": 15.655172413793103, "grad_norm": 27.547733306884766, "learning_rate": 4.156441717791411e-05, "loss": 0.5069, "step": 454 }, { "epoch": 15.724137931034482, "grad_norm": 19.634178161621094, "learning_rate": 4.1257668711656444e-05, "loss": 0.3872, "step": 456 }, { "epoch": 15.793103448275861, "grad_norm": 21.880495071411133, "learning_rate": 4.0950920245398776e-05, "loss": 0.2277, "step": 458 }, { "epoch": 15.862068965517242, "grad_norm": 15.454160690307617, "learning_rate": 4.06441717791411e-05, "loss": 0.3267, "step": 460 }, { "epoch": 15.931034482758621, "grad_norm": 15.292703628540039, "learning_rate": 4.033742331288344e-05, "loss": 0.2948, "step": 462 }, { "epoch": 16.0, "grad_norm": 24.81329917907715, "learning_rate": 4.0030674846625773e-05, "loss": 0.2893, "step": 464 }, { "epoch": 16.0, "eval_accuracy": 0.5984848484848485, "eval_f1_macro": 0.512701663933191, "eval_f1_micro": 0.5984848484848485, "eval_f1_weighted": 0.5832852686300961, "eval_loss": 1.5284953117370605, "eval_precision_macro": 0.5059085452362763, "eval_precision_micro": 0.5984848484848485, "eval_precision_weighted": 0.5752329251259732, "eval_recall_macro": 0.5253136810279667, "eval_recall_micro": 0.5984848484848485, "eval_recall_weighted": 0.5984848484848485, "eval_runtime": 2.1964, "eval_samples_per_second": 60.099, "eval_steps_per_second": 7.74, "step": 464 }, { "epoch": 16.06896551724138, "grad_norm": 19.42568588256836, "learning_rate": 3.97239263803681e-05, "loss": 0.1854, "step": 466 }, { "epoch": 16.137931034482758, "grad_norm": 12.055990219116211, "learning_rate": 3.941717791411043e-05, "loss": 0.2725, "step": 468 }, { "epoch": 16.20689655172414, "grad_norm": 22.306148529052734, "learning_rate": 3.9110429447852764e-05, "loss": 0.2066, "step": 470 }, { "epoch": 16.275862068965516, "grad_norm": 15.890237808227539, "learning_rate": 3.880368098159509e-05, "loss": 0.3127, "step": 472 }, { "epoch": 16.344827586206897, "grad_norm": 17.045835494995117, "learning_rate": 3.849693251533742e-05, "loss": 0.3555, "step": 474 }, { "epoch": 16.413793103448278, "grad_norm": 15.841018676757812, "learning_rate": 3.819018404907976e-05, "loss": 0.3553, "step": 476 }, { "epoch": 16.482758620689655, "grad_norm": 11.002803802490234, "learning_rate": 3.7883435582822086e-05, "loss": 0.3044, "step": 478 }, { "epoch": 16.551724137931036, "grad_norm": 8.885010719299316, "learning_rate": 3.757668711656442e-05, "loss": 0.2396, "step": 480 }, { "epoch": 16.620689655172413, "grad_norm": 20.58298110961914, "learning_rate": 3.726993865030675e-05, "loss": 0.2576, "step": 482 }, { "epoch": 16.689655172413794, "grad_norm": 18.69637107849121, "learning_rate": 3.696319018404908e-05, "loss": 0.3052, "step": 484 }, { "epoch": 16.75862068965517, "grad_norm": 7.023503303527832, "learning_rate": 3.665644171779141e-05, "loss": 0.1449, "step": 486 }, { "epoch": 16.82758620689655, "grad_norm": 18.077198028564453, "learning_rate": 3.634969325153375e-05, "loss": 0.3661, "step": 488 }, { "epoch": 16.896551724137932, "grad_norm": 17.02280616760254, "learning_rate": 3.6042944785276074e-05, "loss": 0.2259, "step": 490 }, { "epoch": 16.96551724137931, "grad_norm": 20.76211929321289, "learning_rate": 3.5736196319018406e-05, "loss": 0.2403, "step": 492 }, { "epoch": 17.0, "eval_accuracy": 0.6287878787878788, "eval_f1_macro": 0.5395048301452242, "eval_f1_micro": 0.6287878787878788, "eval_f1_weighted": 0.6065423604608557, "eval_loss": 1.4820140600204468, "eval_precision_macro": 0.58078231292517, "eval_precision_micro": 0.6287878787878788, "eval_precision_weighted": 0.6380321067821069, "eval_recall_macro": 0.5459561602418744, "eval_recall_micro": 0.6287878787878788, "eval_recall_weighted": 0.6287878787878788, "eval_runtime": 2.1863, "eval_samples_per_second": 60.376, "eval_steps_per_second": 7.776, "step": 493 }, { "epoch": 17.03448275862069, "grad_norm": 20.38401985168457, "learning_rate": 3.542944785276074e-05, "loss": 0.3407, "step": 494 }, { "epoch": 17.103448275862068, "grad_norm": 5.897765636444092, "learning_rate": 3.512269938650307e-05, "loss": 0.1841, "step": 496 }, { "epoch": 17.17241379310345, "grad_norm": 15.455453872680664, "learning_rate": 3.4815950920245396e-05, "loss": 0.1269, "step": 498 }, { "epoch": 17.24137931034483, "grad_norm": 27.898658752441406, "learning_rate": 3.4509202453987735e-05, "loss": 0.2314, "step": 500 }, { "epoch": 17.310344827586206, "grad_norm": 19.497142791748047, "learning_rate": 3.420245398773007e-05, "loss": 0.2743, "step": 502 }, { "epoch": 17.379310344827587, "grad_norm": 15.271653175354004, "learning_rate": 3.3895705521472393e-05, "loss": 0.1961, "step": 504 }, { "epoch": 17.448275862068964, "grad_norm": 25.726585388183594, "learning_rate": 3.3588957055214726e-05, "loss": 0.3415, "step": 506 }, { "epoch": 17.517241379310345, "grad_norm": 22.30027198791504, "learning_rate": 3.328220858895706e-05, "loss": 0.2365, "step": 508 }, { "epoch": 17.586206896551722, "grad_norm": 27.83454132080078, "learning_rate": 3.2975460122699384e-05, "loss": 0.323, "step": 510 }, { "epoch": 17.655172413793103, "grad_norm": 17.994998931884766, "learning_rate": 3.266871165644172e-05, "loss": 0.2189, "step": 512 }, { "epoch": 17.724137931034484, "grad_norm": 15.599747657775879, "learning_rate": 3.2361963190184055e-05, "loss": 0.2235, "step": 514 }, { "epoch": 17.79310344827586, "grad_norm": 19.07125473022461, "learning_rate": 3.205521472392638e-05, "loss": 0.2625, "step": 516 }, { "epoch": 17.862068965517242, "grad_norm": 5.794015407562256, "learning_rate": 3.174846625766871e-05, "loss": 0.1962, "step": 518 }, { "epoch": 17.93103448275862, "grad_norm": 27.05267906188965, "learning_rate": 3.1441717791411045e-05, "loss": 0.4204, "step": 520 }, { "epoch": 18.0, "grad_norm": 12.915875434875488, "learning_rate": 3.113496932515337e-05, "loss": 0.1087, "step": 522 }, { "epoch": 18.0, "eval_accuracy": 0.6060606060606061, "eval_f1_macro": 0.5319546411035773, "eval_f1_micro": 0.6060606060606061, "eval_f1_weighted": 0.6009282162872589, "eval_loss": 1.3999419212341309, "eval_precision_macro": 0.5611877440448869, "eval_precision_micro": 0.6060606060606061, "eval_precision_weighted": 0.6210855415400871, "eval_recall_macro": 0.5260619803476947, "eval_recall_micro": 0.6060606060606061, "eval_recall_weighted": 0.6060606060606061, "eval_runtime": 2.1908, "eval_samples_per_second": 60.251, "eval_steps_per_second": 7.76, "step": 522 }, { "epoch": 18.06896551724138, "grad_norm": 17.82192039489746, "learning_rate": 3.0828220858895703e-05, "loss": 0.1255, "step": 524 }, { "epoch": 18.137931034482758, "grad_norm": 13.459417343139648, "learning_rate": 3.052147239263804e-05, "loss": 0.1728, "step": 526 }, { "epoch": 18.20689655172414, "grad_norm": 19.81383514404297, "learning_rate": 3.0214723926380368e-05, "loss": 0.1743, "step": 528 }, { "epoch": 18.275862068965516, "grad_norm": 17.316072463989258, "learning_rate": 2.99079754601227e-05, "loss": 0.2115, "step": 530 }, { "epoch": 18.344827586206897, "grad_norm": 25.32339096069336, "learning_rate": 2.9601226993865033e-05, "loss": 0.387, "step": 532 }, { "epoch": 18.413793103448278, "grad_norm": 10.883082389831543, "learning_rate": 2.9294478527607362e-05, "loss": 0.0874, "step": 534 }, { "epoch": 18.482758620689655, "grad_norm": 25.40140151977539, "learning_rate": 2.8987730061349694e-05, "loss": 0.3103, "step": 536 }, { "epoch": 18.551724137931036, "grad_norm": 13.151557922363281, "learning_rate": 2.8680981595092026e-05, "loss": 0.277, "step": 538 }, { "epoch": 18.620689655172413, "grad_norm": 16.688093185424805, "learning_rate": 2.837423312883436e-05, "loss": 0.1661, "step": 540 }, { "epoch": 18.689655172413794, "grad_norm": 3.376065492630005, "learning_rate": 2.8067484662576688e-05, "loss": 0.2142, "step": 542 }, { "epoch": 18.75862068965517, "grad_norm": 20.72919464111328, "learning_rate": 2.776073619631902e-05, "loss": 0.2779, "step": 544 }, { "epoch": 18.82758620689655, "grad_norm": 3.795419454574585, "learning_rate": 2.7453987730061353e-05, "loss": 0.0706, "step": 546 }, { "epoch": 18.896551724137932, "grad_norm": 18.60194969177246, "learning_rate": 2.714723926380368e-05, "loss": 0.0951, "step": 548 }, { "epoch": 18.96551724137931, "grad_norm": 17.101329803466797, "learning_rate": 2.6840490797546014e-05, "loss": 0.2619, "step": 550 }, { "epoch": 19.0, "eval_accuracy": 0.6136363636363636, "eval_f1_macro": 0.561813814539783, "eval_f1_micro": 0.6136363636363636, "eval_f1_weighted": 0.6037404914018973, "eval_loss": 1.440819263458252, "eval_precision_macro": 0.6154006028203717, "eval_precision_micro": 0.6136363636363636, "eval_precision_weighted": 0.6224909931745618, "eval_recall_macro": 0.5501284958427816, "eval_recall_micro": 0.6136363636363636, "eval_recall_weighted": 0.6136363636363636, "eval_runtime": 2.201, "eval_samples_per_second": 59.973, "eval_steps_per_second": 7.724, "step": 551 }, { "epoch": 19.03448275862069, "grad_norm": 13.738136291503906, "learning_rate": 2.6533742331288346e-05, "loss": 0.1927, "step": 552 }, { "epoch": 19.103448275862068, "grad_norm": 17.96263885498047, "learning_rate": 2.6226993865030675e-05, "loss": 0.1388, "step": 554 }, { "epoch": 19.17241379310345, "grad_norm": 25.091278076171875, "learning_rate": 2.5920245398773008e-05, "loss": 0.1476, "step": 556 }, { "epoch": 19.24137931034483, "grad_norm": 14.9843168258667, "learning_rate": 2.561349693251534e-05, "loss": 0.116, "step": 558 }, { "epoch": 19.310344827586206, "grad_norm": 13.588825225830078, "learning_rate": 2.530674846625767e-05, "loss": 0.1187, "step": 560 }, { "epoch": 19.379310344827587, "grad_norm": 8.29517650604248, "learning_rate": 2.5e-05, "loss": 0.1494, "step": 562 }, { "epoch": 19.448275862068964, "grad_norm": 24.074113845825195, "learning_rate": 2.469325153374233e-05, "loss": 0.1549, "step": 564 }, { "epoch": 19.517241379310345, "grad_norm": 7.5761213302612305, "learning_rate": 2.4386503067484666e-05, "loss": 0.0937, "step": 566 }, { "epoch": 19.586206896551722, "grad_norm": 9.566593170166016, "learning_rate": 2.4079754601226995e-05, "loss": 0.1075, "step": 568 }, { "epoch": 19.655172413793103, "grad_norm": 14.346840858459473, "learning_rate": 2.3773006134969324e-05, "loss": 0.112, "step": 570 }, { "epoch": 19.724137931034484, "grad_norm": 22.044532775878906, "learning_rate": 2.346625766871166e-05, "loss": 0.1831, "step": 572 }, { "epoch": 19.79310344827586, "grad_norm": 10.464526176452637, "learning_rate": 2.315950920245399e-05, "loss": 0.1353, "step": 574 }, { "epoch": 19.862068965517242, "grad_norm": 6.597527980804443, "learning_rate": 2.285276073619632e-05, "loss": 0.073, "step": 576 }, { "epoch": 19.93103448275862, "grad_norm": 3.7595765590667725, "learning_rate": 2.2546012269938653e-05, "loss": 0.0686, "step": 578 }, { "epoch": 20.0, "grad_norm": 4.094130516052246, "learning_rate": 2.2239263803680982e-05, "loss": 0.1154, "step": 580 }, { "epoch": 20.0, "eval_accuracy": 0.6287878787878788, "eval_f1_macro": 0.5401653994931305, "eval_f1_micro": 0.6287878787878788, "eval_f1_weighted": 0.6089890501655209, "eval_loss": 1.4516006708145142, "eval_precision_macro": 0.5538492063492063, "eval_precision_micro": 0.6287878787878788, "eval_precision_weighted": 0.6144570707070707, "eval_recall_macro": 0.5491912320483749, "eval_recall_micro": 0.6287878787878788, "eval_recall_weighted": 0.6287878787878788, "eval_runtime": 2.2073, "eval_samples_per_second": 59.802, "eval_steps_per_second": 7.702, "step": 580 }, { "epoch": 20.06896551724138, "grad_norm": 9.36839485168457, "learning_rate": 2.1932515337423315e-05, "loss": 0.0687, "step": 582 }, { "epoch": 20.137931034482758, "grad_norm": 24.775922775268555, "learning_rate": 2.1625766871165647e-05, "loss": 0.1334, "step": 584 }, { "epoch": 20.20689655172414, "grad_norm": 23.269336700439453, "learning_rate": 2.1319018404907976e-05, "loss": 0.1813, "step": 586 }, { "epoch": 20.275862068965516, "grad_norm": 5.708396911621094, "learning_rate": 2.1012269938650308e-05, "loss": 0.0841, "step": 588 }, { "epoch": 20.344827586206897, "grad_norm": 11.590498924255371, "learning_rate": 2.0705521472392637e-05, "loss": 0.116, "step": 590 }, { "epoch": 20.413793103448278, "grad_norm": 18.878385543823242, "learning_rate": 2.039877300613497e-05, "loss": 0.1187, "step": 592 }, { "epoch": 20.482758620689655, "grad_norm": 15.631240844726562, "learning_rate": 2.0092024539877302e-05, "loss": 0.1785, "step": 594 }, { "epoch": 20.551724137931036, "grad_norm": 9.58936595916748, "learning_rate": 1.978527607361963e-05, "loss": 0.0336, "step": 596 }, { "epoch": 20.620689655172413, "grad_norm": 15.450642585754395, "learning_rate": 1.9478527607361967e-05, "loss": 0.1695, "step": 598 }, { "epoch": 20.689655172413794, "grad_norm": 4.292616844177246, "learning_rate": 1.9171779141104296e-05, "loss": 0.0633, "step": 600 }, { "epoch": 20.75862068965517, "grad_norm": 4.748676776885986, "learning_rate": 1.8865030674846625e-05, "loss": 0.0889, "step": 602 }, { "epoch": 20.82758620689655, "grad_norm": 16.53461265563965, "learning_rate": 1.855828220858896e-05, "loss": 0.1608, "step": 604 }, { "epoch": 20.896551724137932, "grad_norm": 26.134490966796875, "learning_rate": 1.825153374233129e-05, "loss": 0.1693, "step": 606 }, { "epoch": 20.96551724137931, "grad_norm": 20.084346771240234, "learning_rate": 1.7944785276073618e-05, "loss": 0.1367, "step": 608 }, { "epoch": 21.0, "eval_accuracy": 0.6136363636363636, "eval_f1_macro": 0.5254125588472093, "eval_f1_micro": 0.6136363636363636, "eval_f1_weighted": 0.5942474496633104, "eval_loss": 1.530592441558838, "eval_precision_macro": 0.5321410615528263, "eval_precision_micro": 0.6136363636363636, "eval_precision_weighted": 0.5922574157868276, "eval_recall_macro": 0.5339984882842026, "eval_recall_micro": 0.6136363636363636, "eval_recall_weighted": 0.6136363636363636, "eval_runtime": 2.2038, "eval_samples_per_second": 59.896, "eval_steps_per_second": 7.714, "step": 609 }, { "epoch": 21.03448275862069, "grad_norm": 4.850620746612549, "learning_rate": 1.7638036809815954e-05, "loss": 0.0614, "step": 610 }, { "epoch": 21.103448275862068, "grad_norm": 23.208776473999023, "learning_rate": 1.7331288343558283e-05, "loss": 0.1307, "step": 612 }, { "epoch": 21.17241379310345, "grad_norm": 8.065655708312988, "learning_rate": 1.7024539877300612e-05, "loss": 0.0389, "step": 614 }, { "epoch": 21.24137931034483, "grad_norm": 9.520572662353516, "learning_rate": 1.6717791411042948e-05, "loss": 0.1722, "step": 616 }, { "epoch": 21.310344827586206, "grad_norm": 26.25343894958496, "learning_rate": 1.6411042944785277e-05, "loss": 0.2112, "step": 618 }, { "epoch": 21.379310344827587, "grad_norm": 6.016479015350342, "learning_rate": 1.6104294478527606e-05, "loss": 0.1577, "step": 620 }, { "epoch": 21.448275862068964, "grad_norm": 28.95755386352539, "learning_rate": 1.579754601226994e-05, "loss": 0.1363, "step": 622 }, { "epoch": 21.517241379310345, "grad_norm": 5.472126483917236, "learning_rate": 1.549079754601227e-05, "loss": 0.0527, "step": 624 }, { "epoch": 21.586206896551722, "grad_norm": 12.595699310302734, "learning_rate": 1.5184049079754603e-05, "loss": 0.1044, "step": 626 }, { "epoch": 21.655172413793103, "grad_norm": 13.575519561767578, "learning_rate": 1.4877300613496933e-05, "loss": 0.1545, "step": 628 }, { "epoch": 21.724137931034484, "grad_norm": 1.717926025390625, "learning_rate": 1.4570552147239264e-05, "loss": 0.0754, "step": 630 }, { "epoch": 21.79310344827586, "grad_norm": 15.80093765258789, "learning_rate": 1.4263803680981596e-05, "loss": 0.1765, "step": 632 }, { "epoch": 21.862068965517242, "grad_norm": 8.399004936218262, "learning_rate": 1.3957055214723927e-05, "loss": 0.0478, "step": 634 }, { "epoch": 21.93103448275862, "grad_norm": 9.28877067565918, "learning_rate": 1.3650306748466258e-05, "loss": 0.0561, "step": 636 }, { "epoch": 22.0, "grad_norm": 13.288439750671387, "learning_rate": 1.334355828220859e-05, "loss": 0.0839, "step": 638 }, { "epoch": 22.0, "eval_accuracy": 0.5833333333333334, "eval_f1_macro": 0.515415984334813, "eval_f1_micro": 0.5833333333333334, "eval_f1_weighted": 0.575588621142538, "eval_loss": 1.6396534442901611, "eval_precision_macro": 0.5274346580737558, "eval_precision_micro": 0.5833333333333334, "eval_precision_weighted": 0.5895078605604921, "eval_recall_macro": 0.5252003023431595, "eval_recall_micro": 0.5833333333333334, "eval_recall_weighted": 0.5833333333333334, "eval_runtime": 2.1936, "eval_samples_per_second": 60.176, "eval_steps_per_second": 7.75, "step": 638 }, { "epoch": 22.06896551724138, "grad_norm": 5.867281436920166, "learning_rate": 1.303680981595092e-05, "loss": 0.0999, "step": 640 }, { "epoch": 22.137931034482758, "grad_norm": 3.7638766765594482, "learning_rate": 1.2730061349693251e-05, "loss": 0.0418, "step": 642 }, { "epoch": 22.20689655172414, "grad_norm": 1.9534434080123901, "learning_rate": 1.2423312883435584e-05, "loss": 0.0351, "step": 644 }, { "epoch": 22.275862068965516, "grad_norm": 1.0892353057861328, "learning_rate": 1.2116564417177914e-05, "loss": 0.0272, "step": 646 }, { "epoch": 22.344827586206897, "grad_norm": 16.817415237426758, "learning_rate": 1.1809815950920245e-05, "loss": 0.2253, "step": 648 }, { "epoch": 22.413793103448278, "grad_norm": 24.870695114135742, "learning_rate": 1.1503067484662577e-05, "loss": 0.1316, "step": 650 }, { "epoch": 22.482758620689655, "grad_norm": 22.584014892578125, "learning_rate": 1.119631901840491e-05, "loss": 0.2264, "step": 652 }, { "epoch": 22.551724137931036, "grad_norm": 8.62193775177002, "learning_rate": 1.0889570552147239e-05, "loss": 0.0336, "step": 654 }, { "epoch": 22.620689655172413, "grad_norm": 7.243905067443848, "learning_rate": 1.0582822085889571e-05, "loss": 0.0435, "step": 656 }, { "epoch": 22.689655172413794, "grad_norm": 1.8948745727539062, "learning_rate": 1.0276073619631903e-05, "loss": 0.0655, "step": 658 }, { "epoch": 22.75862068965517, "grad_norm": 5.095564842224121, "learning_rate": 9.969325153374232e-06, "loss": 0.042, "step": 660 }, { "epoch": 22.82758620689655, "grad_norm": 25.0085506439209, "learning_rate": 9.662576687116565e-06, "loss": 0.1186, "step": 662 }, { "epoch": 22.896551724137932, "grad_norm": 4.847318172454834, "learning_rate": 9.355828220858897e-06, "loss": 0.0377, "step": 664 }, { "epoch": 22.96551724137931, "grad_norm": 39.400447845458984, "learning_rate": 9.049079754601228e-06, "loss": 0.1818, "step": 666 }, { "epoch": 23.0, "eval_accuracy": 0.6515151515151515, "eval_f1_macro": 0.565634487061558, "eval_f1_micro": 0.6515151515151515, "eval_f1_weighted": 0.6358752918681777, "eval_loss": 1.641618251800537, "eval_precision_macro": 0.584759718380408, "eval_precision_micro": 0.6515151515151515, "eval_precision_weighted": 0.6455531040170539, "eval_recall_macro": 0.5695616024187452, "eval_recall_micro": 0.6515151515151515, "eval_recall_weighted": 0.6515151515151515, "eval_runtime": 2.2051, "eval_samples_per_second": 59.861, "eval_steps_per_second": 7.709, "step": 667 }, { "epoch": 23.03448275862069, "grad_norm": 11.35802936553955, "learning_rate": 8.742331288343558e-06, "loss": 0.0513, "step": 668 }, { "epoch": 23.103448275862068, "grad_norm": 1.9400774240493774, "learning_rate": 8.435582822085889e-06, "loss": 0.0571, "step": 670 }, { "epoch": 23.17241379310345, "grad_norm": 3.928626775741577, "learning_rate": 8.128834355828221e-06, "loss": 0.034, "step": 672 }, { "epoch": 23.24137931034483, "grad_norm": 5.0317511558532715, "learning_rate": 7.822085889570554e-06, "loss": 0.0833, "step": 674 }, { "epoch": 23.310344827586206, "grad_norm": 12.943672180175781, "learning_rate": 7.5153374233128836e-06, "loss": 0.0575, "step": 676 }, { "epoch": 23.379310344827587, "grad_norm": 1.5223954916000366, "learning_rate": 7.208588957055215e-06, "loss": 0.0237, "step": 678 }, { "epoch": 23.448275862068964, "grad_norm": 21.462011337280273, "learning_rate": 6.901840490797547e-06, "loss": 0.0785, "step": 680 }, { "epoch": 23.517241379310345, "grad_norm": 13.98965072631836, "learning_rate": 6.595092024539877e-06, "loss": 0.0597, "step": 682 }, { "epoch": 23.586206896551722, "grad_norm": 13.07774829864502, "learning_rate": 6.288343558282209e-06, "loss": 0.0498, "step": 684 }, { "epoch": 23.655172413793103, "grad_norm": 2.752511501312256, "learning_rate": 5.98159509202454e-06, "loss": 0.028, "step": 686 }, { "epoch": 23.724137931034484, "grad_norm": 3.4312055110931396, "learning_rate": 5.674846625766871e-06, "loss": 0.0531, "step": 688 }, { "epoch": 23.79310344827586, "grad_norm": 20.920682907104492, "learning_rate": 5.368098159509203e-06, "loss": 0.1892, "step": 690 }, { "epoch": 23.862068965517242, "grad_norm": 1.7730119228363037, "learning_rate": 5.061349693251534e-06, "loss": 0.0409, "step": 692 }, { "epoch": 23.93103448275862, "grad_norm": 8.248014450073242, "learning_rate": 4.7546012269938654e-06, "loss": 0.0499, "step": 694 }, { "epoch": 24.0, "grad_norm": 13.629621505737305, "learning_rate": 4.447852760736196e-06, "loss": 0.0781, "step": 696 }, { "epoch": 24.0, "eval_accuracy": 0.6212121212121212, "eval_f1_macro": 0.5392658545631691, "eval_f1_micro": 0.6212121212121212, "eval_f1_weighted": 0.607896474480917, "eval_loss": 1.6025735139846802, "eval_precision_macro": 0.5523980652552082, "eval_precision_micro": 0.6212121212121212, "eval_precision_weighted": 0.6118077193077193, "eval_recall_macro": 0.5412169312169313, "eval_recall_micro": 0.6212121212121212, "eval_recall_weighted": 0.6212121212121212, "eval_runtime": 2.1909, "eval_samples_per_second": 60.25, "eval_steps_per_second": 7.76, "step": 696 }, { "epoch": 24.06896551724138, "grad_norm": 2.504154920578003, "learning_rate": 4.141104294478528e-06, "loss": 0.0335, "step": 698 }, { "epoch": 24.137931034482758, "grad_norm": 17.774227142333984, "learning_rate": 3.834355828220859e-06, "loss": 0.1496, "step": 700 }, { "epoch": 24.20689655172414, "grad_norm": 2.7488608360290527, "learning_rate": 3.52760736196319e-06, "loss": 0.0266, "step": 702 }, { "epoch": 24.275862068965516, "grad_norm": 7.259423732757568, "learning_rate": 3.2208588957055217e-06, "loss": 0.0863, "step": 704 }, { "epoch": 24.344827586206897, "grad_norm": 7.789644718170166, "learning_rate": 2.914110429447853e-06, "loss": 0.0491, "step": 706 }, { "epoch": 24.413793103448278, "grad_norm": 1.2346043586730957, "learning_rate": 2.607361963190184e-06, "loss": 0.0136, "step": 708 }, { "epoch": 24.482758620689655, "grad_norm": 5.063798904418945, "learning_rate": 2.3006134969325154e-06, "loss": 0.0189, "step": 710 }, { "epoch": 24.551724137931036, "grad_norm": 6.230537414550781, "learning_rate": 1.9938650306748465e-06, "loss": 0.0431, "step": 712 }, { "epoch": 24.620689655172413, "grad_norm": 15.741507530212402, "learning_rate": 1.687116564417178e-06, "loss": 0.19, "step": 714 }, { "epoch": 24.689655172413794, "grad_norm": 12.17914867401123, "learning_rate": 1.3803680981595093e-06, "loss": 0.0662, "step": 716 }, { "epoch": 24.75862068965517, "grad_norm": 1.5245462656021118, "learning_rate": 1.0736196319018406e-06, "loss": 0.0579, "step": 718 }, { "epoch": 24.82758620689655, "grad_norm": 3.734255790710449, "learning_rate": 7.668711656441718e-07, "loss": 0.0337, "step": 720 }, { "epoch": 24.896551724137932, "grad_norm": 25.646230697631836, "learning_rate": 4.601226993865031e-07, "loss": 0.1464, "step": 722 }, { "epoch": 24.96551724137931, "grad_norm": 11.875574111938477, "learning_rate": 1.5337423312883438e-07, "loss": 0.0792, "step": 724 }, { "epoch": 25.0, "eval_accuracy": 0.6287878787878788, "eval_f1_macro": 0.5494495726426264, "eval_f1_micro": 0.6287878787878788, "eval_f1_weighted": 0.6179503958679277, "eval_loss": 1.599715232849121, "eval_precision_macro": 0.5716202716202715, "eval_precision_micro": 0.6287878787878788, "eval_precision_weighted": 0.6297404683768321, "eval_recall_macro": 0.5480196523053665, "eval_recall_micro": 0.6287878787878788, "eval_recall_weighted": 0.6287878787878788, "eval_runtime": 2.2024, "eval_samples_per_second": 59.933, "eval_steps_per_second": 7.719, "step": 725 }, { "epoch": 25.0, "step": 725, "total_flos": 5.76425379898368e+16, "train_loss": 0.7207291752639515, "train_runtime": 641.8354, "train_samples_per_second": 17.995, "train_steps_per_second": 1.13 } ], "logging_steps": 2, "max_steps": 725, "num_input_tokens_seen": 0, "num_train_epochs": 25, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.76425379898368e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }