| { | |
| "best_metric": 1.1496680974960327, | |
| "best_model_checkpoint": "square_run_min_loss/checkpoint-986", | |
| "epoch": 35.0, | |
| "eval_steps": 500, | |
| "global_step": 2030, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.034482758620689655, | |
| "grad_norm": 2.575118064880371, | |
| "learning_rate": 9.852216748768474e-07, | |
| "loss": 1.9523, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.06896551724137931, | |
| "grad_norm": 4.115417003631592, | |
| "learning_rate": 1.970443349753695e-06, | |
| "loss": 1.897, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.10344827586206896, | |
| "grad_norm": 2.1270015239715576, | |
| "learning_rate": 2.955665024630542e-06, | |
| "loss": 1.9272, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.13793103448275862, | |
| "grad_norm": 3.1353254318237305, | |
| "learning_rate": 3.94088669950739e-06, | |
| "loss": 1.9155, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.1724137931034483, | |
| "grad_norm": 2.0807693004608154, | |
| "learning_rate": 4.926108374384237e-06, | |
| "loss": 1.9557, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.20689655172413793, | |
| "grad_norm": 2.7346973419189453, | |
| "learning_rate": 5.911330049261084e-06, | |
| "loss": 1.9696, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.2413793103448276, | |
| "grad_norm": 2.41186261177063, | |
| "learning_rate": 6.896551724137932e-06, | |
| "loss": 1.927, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.27586206896551724, | |
| "grad_norm": 3.0956790447235107, | |
| "learning_rate": 7.88177339901478e-06, | |
| "loss": 1.9232, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.3103448275862069, | |
| "grad_norm": 2.850667953491211, | |
| "learning_rate": 8.866995073891627e-06, | |
| "loss": 1.9046, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.3448275862068966, | |
| "grad_norm": 2.605441093444824, | |
| "learning_rate": 9.852216748768475e-06, | |
| "loss": 1.904, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.3793103448275862, | |
| "grad_norm": 2.4900379180908203, | |
| "learning_rate": 1.0837438423645322e-05, | |
| "loss": 1.8945, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.41379310344827586, | |
| "grad_norm": 2.4673266410827637, | |
| "learning_rate": 1.1822660098522168e-05, | |
| "loss": 2.0077, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.4482758620689655, | |
| "grad_norm": 2.4353814125061035, | |
| "learning_rate": 1.2807881773399016e-05, | |
| "loss": 2.0005, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.4827586206896552, | |
| "grad_norm": 2.159104824066162, | |
| "learning_rate": 1.3793103448275863e-05, | |
| "loss": 1.8958, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.5172413793103449, | |
| "grad_norm": 2.747079849243164, | |
| "learning_rate": 1.4778325123152711e-05, | |
| "loss": 2.0059, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.5517241379310345, | |
| "grad_norm": 1.8319206237792969, | |
| "learning_rate": 1.576354679802956e-05, | |
| "loss": 1.8849, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.5862068965517241, | |
| "grad_norm": 2.87062668800354, | |
| "learning_rate": 1.6748768472906403e-05, | |
| "loss": 1.8649, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.6206896551724138, | |
| "grad_norm": 3.053698778152466, | |
| "learning_rate": 1.7733990147783254e-05, | |
| "loss": 1.8314, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.6551724137931034, | |
| "grad_norm": 2.7139945030212402, | |
| "learning_rate": 1.8719211822660098e-05, | |
| "loss": 1.8343, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.6896551724137931, | |
| "grad_norm": 2.85752534866333, | |
| "learning_rate": 1.970443349753695e-05, | |
| "loss": 1.9139, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.7241379310344828, | |
| "grad_norm": 2.415708303451538, | |
| "learning_rate": 2.0689655172413793e-05, | |
| "loss": 1.8513, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.7586206896551724, | |
| "grad_norm": 2.8618996143341064, | |
| "learning_rate": 2.1674876847290644e-05, | |
| "loss": 1.9395, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.7931034482758621, | |
| "grad_norm": 2.4992380142211914, | |
| "learning_rate": 2.266009852216749e-05, | |
| "loss": 1.9028, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.8275862068965517, | |
| "grad_norm": 1.5736284255981445, | |
| "learning_rate": 2.3645320197044336e-05, | |
| "loss": 1.8803, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.8620689655172413, | |
| "grad_norm": 2.486845016479492, | |
| "learning_rate": 2.4630541871921184e-05, | |
| "loss": 1.8354, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.896551724137931, | |
| "grad_norm": 2.764396905899048, | |
| "learning_rate": 2.561576354679803e-05, | |
| "loss": 1.8192, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.9310344827586207, | |
| "grad_norm": 1.9337031841278076, | |
| "learning_rate": 2.660098522167488e-05, | |
| "loss": 1.9035, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.9655172413793104, | |
| "grad_norm": 3.11140775680542, | |
| "learning_rate": 2.7586206896551727e-05, | |
| "loss": 1.9185, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 2.8587818145751953, | |
| "learning_rate": 2.857142857142857e-05, | |
| "loss": 1.934, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.20454545454545456, | |
| "eval_f1_macro": 0.06638977430052481, | |
| "eval_f1_micro": 0.20454545454545456, | |
| "eval_f1_weighted": 0.09010661039058604, | |
| "eval_loss": 1.877989649772644, | |
| "eval_precision_macro": 0.17075892857142858, | |
| "eval_precision_micro": 0.20454545454545456, | |
| "eval_precision_weighted": 0.24153645833333334, | |
| "eval_recall_macro": 0.15343915343915343, | |
| "eval_recall_micro": 0.20454545454545456, | |
| "eval_recall_weighted": 0.20454545454545456, | |
| "eval_runtime": 1.8422, | |
| "eval_samples_per_second": 71.652, | |
| "eval_steps_per_second": 9.228, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.0344827586206897, | |
| "grad_norm": 2.1993751525878906, | |
| "learning_rate": 2.9556650246305422e-05, | |
| "loss": 1.8763, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.0689655172413792, | |
| "grad_norm": 3.214564561843872, | |
| "learning_rate": 3.0541871921182266e-05, | |
| "loss": 1.9158, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.103448275862069, | |
| "grad_norm": 2.2450172901153564, | |
| "learning_rate": 3.152709359605912e-05, | |
| "loss": 1.7313, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.1379310344827587, | |
| "grad_norm": 3.495370388031006, | |
| "learning_rate": 3.251231527093596e-05, | |
| "loss": 1.77, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.1724137931034484, | |
| "grad_norm": 2.372056007385254, | |
| "learning_rate": 3.3497536945812806e-05, | |
| "loss": 1.8228, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.206896551724138, | |
| "grad_norm": 2.4576425552368164, | |
| "learning_rate": 3.4482758620689657e-05, | |
| "loss": 1.9171, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.2413793103448276, | |
| "grad_norm": 2.064105987548828, | |
| "learning_rate": 3.546798029556651e-05, | |
| "loss": 1.9479, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.2758620689655173, | |
| "grad_norm": 1.6881994009017944, | |
| "learning_rate": 3.645320197044335e-05, | |
| "loss": 2.0201, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.3103448275862069, | |
| "grad_norm": 2.1131479740142822, | |
| "learning_rate": 3.7438423645320196e-05, | |
| "loss": 1.7073, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.3448275862068966, | |
| "grad_norm": 2.1655850410461426, | |
| "learning_rate": 3.842364532019704e-05, | |
| "loss": 1.8294, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.3793103448275863, | |
| "grad_norm": 2.9730122089385986, | |
| "learning_rate": 3.94088669950739e-05, | |
| "loss": 1.9662, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.4137931034482758, | |
| "grad_norm": 2.5296716690063477, | |
| "learning_rate": 4.039408866995074e-05, | |
| "loss": 1.8582, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.4482758620689655, | |
| "grad_norm": 2.682340145111084, | |
| "learning_rate": 4.1379310344827587e-05, | |
| "loss": 1.9321, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.4827586206896552, | |
| "grad_norm": 2.764066457748413, | |
| "learning_rate": 4.236453201970443e-05, | |
| "loss": 1.8595, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.5172413793103448, | |
| "grad_norm": 3.0940446853637695, | |
| "learning_rate": 4.334975369458129e-05, | |
| "loss": 2.0041, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.5517241379310345, | |
| "grad_norm": 2.3535633087158203, | |
| "learning_rate": 4.433497536945813e-05, | |
| "loss": 2.0397, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.5862068965517242, | |
| "grad_norm": 2.1643190383911133, | |
| "learning_rate": 4.532019704433498e-05, | |
| "loss": 1.8423, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.6206896551724137, | |
| "grad_norm": 2.2240335941314697, | |
| "learning_rate": 4.630541871921182e-05, | |
| "loss": 1.8856, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.6551724137931034, | |
| "grad_norm": 1.8210374116897583, | |
| "learning_rate": 4.729064039408867e-05, | |
| "loss": 1.8442, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.6896551724137931, | |
| "grad_norm": 2.3819000720977783, | |
| "learning_rate": 4.827586206896552e-05, | |
| "loss": 1.9759, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.7241379310344827, | |
| "grad_norm": 2.209657907485962, | |
| "learning_rate": 4.926108374384237e-05, | |
| "loss": 1.8977, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.7586206896551724, | |
| "grad_norm": 1.8794820308685303, | |
| "learning_rate": 5.024630541871922e-05, | |
| "loss": 1.8373, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.793103448275862, | |
| "grad_norm": 1.9075384140014648, | |
| "learning_rate": 5.123152709359606e-05, | |
| "loss": 1.9054, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.8275862068965516, | |
| "grad_norm": 2.074863910675049, | |
| "learning_rate": 5.2216748768472914e-05, | |
| "loss": 1.9374, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.8620689655172413, | |
| "grad_norm": 3.287247657775879, | |
| "learning_rate": 5.320197044334976e-05, | |
| "loss": 1.9644, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.896551724137931, | |
| "grad_norm": 2.7543118000030518, | |
| "learning_rate": 5.41871921182266e-05, | |
| "loss": 1.8498, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.9310344827586206, | |
| "grad_norm": 2.3745720386505127, | |
| "learning_rate": 5.517241379310345e-05, | |
| "loss": 1.8688, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.9655172413793105, | |
| "grad_norm": 1.6657403707504272, | |
| "learning_rate": 5.61576354679803e-05, | |
| "loss": 1.8744, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 1.9578406810760498, | |
| "learning_rate": 5.714285714285714e-05, | |
| "loss": 1.8145, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.17424242424242425, | |
| "eval_f1_macro": 0.06911375661375661, | |
| "eval_f1_micro": 0.17424242424242425, | |
| "eval_f1_weighted": 0.07554713804713804, | |
| "eval_loss": 1.882804274559021, | |
| "eval_precision_macro": 0.06077694235588972, | |
| "eval_precision_micro": 0.17424242424242425, | |
| "eval_precision_weighted": 0.06578947368421052, | |
| "eval_recall_macro": 0.1574829931972789, | |
| "eval_recall_micro": 0.17424242424242425, | |
| "eval_recall_weighted": 0.17424242424242425, | |
| "eval_runtime": 1.6622, | |
| "eval_samples_per_second": 79.414, | |
| "eval_steps_per_second": 10.228, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.0344827586206895, | |
| "grad_norm": 3.171048879623413, | |
| "learning_rate": 5.8128078817733986e-05, | |
| "loss": 1.9026, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 2.0689655172413794, | |
| "grad_norm": 2.3211982250213623, | |
| "learning_rate": 5.9113300492610844e-05, | |
| "loss": 1.9666, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.103448275862069, | |
| "grad_norm": 2.0532777309417725, | |
| "learning_rate": 6.0098522167487695e-05, | |
| "loss": 1.8683, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.1379310344827585, | |
| "grad_norm": 2.8642830848693848, | |
| "learning_rate": 6.108374384236453e-05, | |
| "loss": 1.8174, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.1724137931034484, | |
| "grad_norm": 2.02150821685791, | |
| "learning_rate": 6.206896551724138e-05, | |
| "loss": 1.837, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.206896551724138, | |
| "grad_norm": 2.0854902267456055, | |
| "learning_rate": 6.305418719211823e-05, | |
| "loss": 1.8358, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 2.2413793103448274, | |
| "grad_norm": 2.257364511489868, | |
| "learning_rate": 6.403940886699507e-05, | |
| "loss": 1.7473, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.2758620689655173, | |
| "grad_norm": 2.777860641479492, | |
| "learning_rate": 6.502463054187192e-05, | |
| "loss": 2.0573, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 2.310344827586207, | |
| "grad_norm": 1.8564225435256958, | |
| "learning_rate": 6.600985221674877e-05, | |
| "loss": 1.7209, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 2.344827586206897, | |
| "grad_norm": 1.9192124605178833, | |
| "learning_rate": 6.699507389162561e-05, | |
| "loss": 1.8069, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 2.3793103448275863, | |
| "grad_norm": 2.5344839096069336, | |
| "learning_rate": 6.798029556650246e-05, | |
| "loss": 1.9476, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 2.413793103448276, | |
| "grad_norm": 2.2182462215423584, | |
| "learning_rate": 6.896551724137931e-05, | |
| "loss": 1.7292, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.4482758620689653, | |
| "grad_norm": 3.1175408363342285, | |
| "learning_rate": 6.995073891625616e-05, | |
| "loss": 1.8131, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 2.4827586206896552, | |
| "grad_norm": 2.5868771076202393, | |
| "learning_rate": 7.093596059113302e-05, | |
| "loss": 1.8445, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 2.5172413793103448, | |
| "grad_norm": 2.509629726409912, | |
| "learning_rate": 7.192118226600985e-05, | |
| "loss": 1.7201, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 2.5517241379310347, | |
| "grad_norm": 1.862570881843567, | |
| "learning_rate": 7.29064039408867e-05, | |
| "loss": 1.6859, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 2.586206896551724, | |
| "grad_norm": 2.31773042678833, | |
| "learning_rate": 7.389162561576355e-05, | |
| "loss": 1.754, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.6206896551724137, | |
| "grad_norm": 2.858905076980591, | |
| "learning_rate": 7.487684729064039e-05, | |
| "loss": 1.8169, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 2.655172413793103, | |
| "grad_norm": 4.044278621673584, | |
| "learning_rate": 7.586206896551724e-05, | |
| "loss": 1.871, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 2.689655172413793, | |
| "grad_norm": 2.7265326976776123, | |
| "learning_rate": 7.684729064039408e-05, | |
| "loss": 1.8314, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 2.7241379310344827, | |
| "grad_norm": 2.775766611099243, | |
| "learning_rate": 7.783251231527095e-05, | |
| "loss": 1.8483, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 2.7586206896551726, | |
| "grad_norm": 3.1784214973449707, | |
| "learning_rate": 7.88177339901478e-05, | |
| "loss": 1.7737, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.793103448275862, | |
| "grad_norm": 2.7301290035247803, | |
| "learning_rate": 7.980295566502463e-05, | |
| "loss": 1.8225, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 2.8275862068965516, | |
| "grad_norm": 3.0388286113739014, | |
| "learning_rate": 8.078817733990148e-05, | |
| "loss": 1.7341, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 2.862068965517241, | |
| "grad_norm": 2.806058883666992, | |
| "learning_rate": 8.177339901477834e-05, | |
| "loss": 1.9648, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 2.896551724137931, | |
| "grad_norm": 1.8368113040924072, | |
| "learning_rate": 8.275862068965517e-05, | |
| "loss": 1.7451, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 2.9310344827586206, | |
| "grad_norm": 2.3558056354522705, | |
| "learning_rate": 8.374384236453202e-05, | |
| "loss": 1.8633, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.9655172413793105, | |
| "grad_norm": 3.564084053039551, | |
| "learning_rate": 8.472906403940886e-05, | |
| "loss": 1.8561, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 2.9199440479278564, | |
| "learning_rate": 8.571428571428571e-05, | |
| "loss": 1.8527, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.3787878787878788, | |
| "eval_f1_macro": 0.25029652886795745, | |
| "eval_f1_micro": 0.3787878787878788, | |
| "eval_f1_weighted": 0.30532939282939286, | |
| "eval_loss": 1.7130813598632812, | |
| "eval_precision_macro": 0.25725108225108223, | |
| "eval_precision_micro": 0.3787878787878788, | |
| "eval_precision_weighted": 0.30621556473829203, | |
| "eval_recall_macro": 0.3094028722600151, | |
| "eval_recall_micro": 0.3787878787878788, | |
| "eval_recall_weighted": 0.3787878787878788, | |
| "eval_runtime": 1.75, | |
| "eval_samples_per_second": 75.429, | |
| "eval_steps_per_second": 9.714, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 3.0344827586206895, | |
| "grad_norm": 2.774258613586426, | |
| "learning_rate": 8.669950738916258e-05, | |
| "loss": 1.8599, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 3.0689655172413794, | |
| "grad_norm": 2.9602231979370117, | |
| "learning_rate": 8.768472906403941e-05, | |
| "loss": 1.736, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 3.103448275862069, | |
| "grad_norm": 2.8377699851989746, | |
| "learning_rate": 8.866995073891627e-05, | |
| "loss": 1.5466, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 3.1379310344827585, | |
| "grad_norm": 2.192164182662964, | |
| "learning_rate": 8.96551724137931e-05, | |
| "loss": 1.7214, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 3.1724137931034484, | |
| "grad_norm": 3.274397134780884, | |
| "learning_rate": 9.064039408866995e-05, | |
| "loss": 1.7141, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 3.206896551724138, | |
| "grad_norm": 2.574033737182617, | |
| "learning_rate": 9.16256157635468e-05, | |
| "loss": 1.5941, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 3.2413793103448274, | |
| "grad_norm": 2.457569122314453, | |
| "learning_rate": 9.261083743842364e-05, | |
| "loss": 1.7504, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 3.2758620689655173, | |
| "grad_norm": 3.9681930541992188, | |
| "learning_rate": 9.35960591133005e-05, | |
| "loss": 1.7611, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 3.310344827586207, | |
| "grad_norm": 3.0164175033569336, | |
| "learning_rate": 9.458128078817734e-05, | |
| "loss": 1.8015, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 3.344827586206897, | |
| "grad_norm": 2.874098539352417, | |
| "learning_rate": 9.55665024630542e-05, | |
| "loss": 1.7201, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 3.3793103448275863, | |
| "grad_norm": 2.657230854034424, | |
| "learning_rate": 9.655172413793105e-05, | |
| "loss": 1.7127, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 3.413793103448276, | |
| "grad_norm": 3.1468231678009033, | |
| "learning_rate": 9.753694581280788e-05, | |
| "loss": 1.6399, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 3.4482758620689653, | |
| "grad_norm": 2.7480523586273193, | |
| "learning_rate": 9.852216748768474e-05, | |
| "loss": 1.5484, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 3.4827586206896552, | |
| "grad_norm": 3.981072187423706, | |
| "learning_rate": 9.950738916256159e-05, | |
| "loss": 1.9724, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 3.5172413793103448, | |
| "grad_norm": 3.0168025493621826, | |
| "learning_rate": 9.994526546250684e-05, | |
| "loss": 1.7675, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 3.5517241379310347, | |
| "grad_norm": 3.0766713619232178, | |
| "learning_rate": 9.983579638752053e-05, | |
| "loss": 1.7034, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 3.586206896551724, | |
| "grad_norm": 2.499846935272217, | |
| "learning_rate": 9.972632731253421e-05, | |
| "loss": 1.7265, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 3.6206896551724137, | |
| "grad_norm": 2.9722824096679688, | |
| "learning_rate": 9.96168582375479e-05, | |
| "loss": 1.7595, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 3.655172413793103, | |
| "grad_norm": 3.6022815704345703, | |
| "learning_rate": 9.950738916256159e-05, | |
| "loss": 2.0972, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 3.689655172413793, | |
| "grad_norm": 2.596907377243042, | |
| "learning_rate": 9.939792008757527e-05, | |
| "loss": 1.6429, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 3.7241379310344827, | |
| "grad_norm": 4.036109924316406, | |
| "learning_rate": 9.928845101258894e-05, | |
| "loss": 1.7363, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 3.7586206896551726, | |
| "grad_norm": 3.0401430130004883, | |
| "learning_rate": 9.917898193760263e-05, | |
| "loss": 1.7639, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 3.793103448275862, | |
| "grad_norm": 3.6351616382598877, | |
| "learning_rate": 9.906951286261632e-05, | |
| "loss": 1.8102, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 3.8275862068965516, | |
| "grad_norm": 1.9515084028244019, | |
| "learning_rate": 9.896004378762999e-05, | |
| "loss": 1.6084, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 3.862068965517241, | |
| "grad_norm": 3.5374197959899902, | |
| "learning_rate": 9.885057471264369e-05, | |
| "loss": 2.0067, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 3.896551724137931, | |
| "grad_norm": 3.3599019050598145, | |
| "learning_rate": 9.874110563765738e-05, | |
| "loss": 1.7673, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 3.9310344827586206, | |
| "grad_norm": 2.432833194732666, | |
| "learning_rate": 9.863163656267105e-05, | |
| "loss": 1.9012, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 3.9655172413793105, | |
| "grad_norm": 1.9806126356124878, | |
| "learning_rate": 9.852216748768474e-05, | |
| "loss": 1.862, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 3.215442419052124, | |
| "learning_rate": 9.841269841269841e-05, | |
| "loss": 1.6734, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_accuracy": 0.2803030303030303, | |
| "eval_f1_macro": 0.1620519748179323, | |
| "eval_f1_micro": 0.2803030303030303, | |
| "eval_f1_weighted": 0.20874649385287686, | |
| "eval_loss": 1.794013261795044, | |
| "eval_precision_macro": 0.21449586012692806, | |
| "eval_precision_micro": 0.2803030303030303, | |
| "eval_precision_weighted": 0.2623532322340795, | |
| "eval_recall_macro": 0.2075585789871504, | |
| "eval_recall_micro": 0.2803030303030303, | |
| "eval_recall_weighted": 0.2803030303030303, | |
| "eval_runtime": 1.7188, | |
| "eval_samples_per_second": 76.796, | |
| "eval_steps_per_second": 9.89, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 4.0344827586206895, | |
| "grad_norm": 1.8983070850372314, | |
| "learning_rate": 9.83032293377121e-05, | |
| "loss": 1.8275, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 4.068965517241379, | |
| "grad_norm": 2.512375593185425, | |
| "learning_rate": 9.819376026272578e-05, | |
| "loss": 1.8915, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 4.103448275862069, | |
| "grad_norm": 2.564410924911499, | |
| "learning_rate": 9.808429118773947e-05, | |
| "loss": 1.6557, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 4.137931034482759, | |
| "grad_norm": 4.520391464233398, | |
| "learning_rate": 9.797482211275315e-05, | |
| "loss": 1.6821, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 4.172413793103448, | |
| "grad_norm": 2.6028499603271484, | |
| "learning_rate": 9.786535303776684e-05, | |
| "loss": 1.5911, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 4.206896551724138, | |
| "grad_norm": 2.6997692584991455, | |
| "learning_rate": 9.775588396278051e-05, | |
| "loss": 1.6449, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 4.241379310344827, | |
| "grad_norm": 5.493044376373291, | |
| "learning_rate": 9.76464148877942e-05, | |
| "loss": 1.6607, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 4.275862068965517, | |
| "grad_norm": 3.518596887588501, | |
| "learning_rate": 9.753694581280788e-05, | |
| "loss": 1.7582, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 4.310344827586207, | |
| "grad_norm": 5.089727401733398, | |
| "learning_rate": 9.742747673782157e-05, | |
| "loss": 1.64, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 4.344827586206897, | |
| "grad_norm": 2.1755893230438232, | |
| "learning_rate": 9.731800766283526e-05, | |
| "loss": 1.7475, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 4.379310344827586, | |
| "grad_norm": 3.20668363571167, | |
| "learning_rate": 9.720853858784894e-05, | |
| "loss": 1.683, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 4.413793103448276, | |
| "grad_norm": 1.785972237586975, | |
| "learning_rate": 9.709906951286262e-05, | |
| "loss": 1.6876, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 4.448275862068965, | |
| "grad_norm": 3.6471712589263916, | |
| "learning_rate": 9.69896004378763e-05, | |
| "loss": 2.1599, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 4.482758620689655, | |
| "grad_norm": 2.0776071548461914, | |
| "learning_rate": 9.688013136288999e-05, | |
| "loss": 2.009, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 4.517241379310345, | |
| "grad_norm": 2.757769823074341, | |
| "learning_rate": 9.677066228790367e-05, | |
| "loss": 2.0974, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 4.551724137931035, | |
| "grad_norm": 2.731898307800293, | |
| "learning_rate": 9.666119321291736e-05, | |
| "loss": 1.4793, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 4.586206896551724, | |
| "grad_norm": 2.585442543029785, | |
| "learning_rate": 9.655172413793105e-05, | |
| "loss": 1.8863, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 4.620689655172414, | |
| "grad_norm": 2.7498419284820557, | |
| "learning_rate": 9.644225506294472e-05, | |
| "loss": 1.6424, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 4.655172413793103, | |
| "grad_norm": 4.065057754516602, | |
| "learning_rate": 9.63327859879584e-05, | |
| "loss": 1.5362, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 4.689655172413794, | |
| "grad_norm": 3.4030730724334717, | |
| "learning_rate": 9.622331691297209e-05, | |
| "loss": 1.7119, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 4.724137931034483, | |
| "grad_norm": 3.3285632133483887, | |
| "learning_rate": 9.611384783798576e-05, | |
| "loss": 1.7643, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 4.758620689655173, | |
| "grad_norm": 3.027642011642456, | |
| "learning_rate": 9.600437876299946e-05, | |
| "loss": 1.7388, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 4.793103448275862, | |
| "grad_norm": 2.7076125144958496, | |
| "learning_rate": 9.589490968801315e-05, | |
| "loss": 1.7675, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 4.827586206896552, | |
| "grad_norm": 3.127922534942627, | |
| "learning_rate": 9.578544061302682e-05, | |
| "loss": 1.4597, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 4.862068965517241, | |
| "grad_norm": 4.732183456420898, | |
| "learning_rate": 9.567597153804051e-05, | |
| "loss": 1.7048, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 4.896551724137931, | |
| "grad_norm": 4.8749284744262695, | |
| "learning_rate": 9.55665024630542e-05, | |
| "loss": 1.8464, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 4.931034482758621, | |
| "grad_norm": 2.5762221813201904, | |
| "learning_rate": 9.545703338806787e-05, | |
| "loss": 1.9569, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 4.9655172413793105, | |
| "grad_norm": 3.143484354019165, | |
| "learning_rate": 9.534756431308155e-05, | |
| "loss": 1.7673, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 2.596794843673706, | |
| "learning_rate": 9.523809523809524e-05, | |
| "loss": 1.6408, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_accuracy": 0.3333333333333333, | |
| "eval_f1_macro": 0.156998556998557, | |
| "eval_f1_micro": 0.3333333333333333, | |
| "eval_f1_weighted": 0.19651056014692378, | |
| "eval_loss": 1.6808322668075562, | |
| "eval_precision_macro": 0.14322344322344321, | |
| "eval_precision_micro": 0.3333333333333333, | |
| "eval_precision_weighted": 0.18583916083916083, | |
| "eval_recall_macro": 0.27017384731670446, | |
| "eval_recall_micro": 0.3333333333333333, | |
| "eval_recall_weighted": 0.3333333333333333, | |
| "eval_runtime": 1.7386, | |
| "eval_samples_per_second": 75.923, | |
| "eval_steps_per_second": 9.778, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 5.0344827586206895, | |
| "grad_norm": 3.2708096504211426, | |
| "learning_rate": 9.512862616310893e-05, | |
| "loss": 2.1016, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 5.068965517241379, | |
| "grad_norm": 2.2087342739105225, | |
| "learning_rate": 9.501915708812261e-05, | |
| "loss": 1.3591, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 5.103448275862069, | |
| "grad_norm": 3.294851779937744, | |
| "learning_rate": 9.490968801313629e-05, | |
| "loss": 1.8157, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 5.137931034482759, | |
| "grad_norm": 2.4422295093536377, | |
| "learning_rate": 9.480021893814997e-05, | |
| "loss": 1.4012, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 5.172413793103448, | |
| "grad_norm": 2.5836057662963867, | |
| "learning_rate": 9.469074986316366e-05, | |
| "loss": 1.6883, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 5.206896551724138, | |
| "grad_norm": 3.3566815853118896, | |
| "learning_rate": 9.458128078817734e-05, | |
| "loss": 1.4683, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 5.241379310344827, | |
| "grad_norm": 2.7103140354156494, | |
| "learning_rate": 9.447181171319103e-05, | |
| "loss": 1.7214, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 5.275862068965517, | |
| "grad_norm": 1.974759578704834, | |
| "learning_rate": 9.436234263820472e-05, | |
| "loss": 1.7763, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 5.310344827586207, | |
| "grad_norm": 3.063124418258667, | |
| "learning_rate": 9.425287356321839e-05, | |
| "loss": 1.6032, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 5.344827586206897, | |
| "grad_norm": 2.366842269897461, | |
| "learning_rate": 9.414340448823208e-05, | |
| "loss": 1.771, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 5.379310344827586, | |
| "grad_norm": 4.095522403717041, | |
| "learning_rate": 9.403393541324576e-05, | |
| "loss": 1.6587, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 5.413793103448276, | |
| "grad_norm": 3.9357569217681885, | |
| "learning_rate": 9.392446633825945e-05, | |
| "loss": 1.3663, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 5.448275862068965, | |
| "grad_norm": 4.434700965881348, | |
| "learning_rate": 9.381499726327313e-05, | |
| "loss": 1.5951, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 5.482758620689655, | |
| "grad_norm": 2.257758855819702, | |
| "learning_rate": 9.370552818828682e-05, | |
| "loss": 1.3281, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 5.517241379310345, | |
| "grad_norm": 2.474900960922241, | |
| "learning_rate": 9.35960591133005e-05, | |
| "loss": 1.3842, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 5.551724137931035, | |
| "grad_norm": 2.744278907775879, | |
| "learning_rate": 9.348659003831418e-05, | |
| "loss": 1.6494, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 5.586206896551724, | |
| "grad_norm": 3.2747697830200195, | |
| "learning_rate": 9.337712096332787e-05, | |
| "loss": 1.3476, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 5.620689655172414, | |
| "grad_norm": 3.0323941707611084, | |
| "learning_rate": 9.326765188834154e-05, | |
| "loss": 1.3379, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 5.655172413793103, | |
| "grad_norm": 2.5281591415405273, | |
| "learning_rate": 9.315818281335524e-05, | |
| "loss": 1.4997, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 5.689655172413794, | |
| "grad_norm": 4.624899864196777, | |
| "learning_rate": 9.304871373836892e-05, | |
| "loss": 1.5624, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 5.724137931034483, | |
| "grad_norm": 2.9435858726501465, | |
| "learning_rate": 9.29392446633826e-05, | |
| "loss": 1.3056, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 5.758620689655173, | |
| "grad_norm": 2.5763392448425293, | |
| "learning_rate": 9.282977558839628e-05, | |
| "loss": 1.5451, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 5.793103448275862, | |
| "grad_norm": 5.08242654800415, | |
| "learning_rate": 9.272030651340997e-05, | |
| "loss": 1.7072, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 5.827586206896552, | |
| "grad_norm": 2.7309770584106445, | |
| "learning_rate": 9.261083743842364e-05, | |
| "loss": 1.2524, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 5.862068965517241, | |
| "grad_norm": 5.227554798126221, | |
| "learning_rate": 9.250136836343733e-05, | |
| "loss": 1.5564, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 5.896551724137931, | |
| "grad_norm": 2.266554117202759, | |
| "learning_rate": 9.239189928845102e-05, | |
| "loss": 1.6249, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 5.931034482758621, | |
| "grad_norm": 3.1214962005615234, | |
| "learning_rate": 9.22824302134647e-05, | |
| "loss": 1.7091, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 5.9655172413793105, | |
| "grad_norm": 2.7962958812713623, | |
| "learning_rate": 9.217296113847839e-05, | |
| "loss": 1.7433, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 3.0703368186950684, | |
| "learning_rate": 9.206349206349206e-05, | |
| "loss": 1.5696, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_accuracy": 0.44696969696969696, | |
| "eval_f1_macro": 0.31719822149201493, | |
| "eval_f1_micro": 0.44696969696969696, | |
| "eval_f1_weighted": 0.38023986322449504, | |
| "eval_loss": 1.5061465501785278, | |
| "eval_precision_macro": 0.38953973562669214, | |
| "eval_precision_micro": 0.44696969696969696, | |
| "eval_precision_weighted": 0.41856035269078745, | |
| "eval_recall_macro": 0.3617989417989418, | |
| "eval_recall_micro": 0.44696969696969696, | |
| "eval_recall_weighted": 0.44696969696969696, | |
| "eval_runtime": 1.7478, | |
| "eval_samples_per_second": 75.524, | |
| "eval_steps_per_second": 9.727, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 6.0344827586206895, | |
| "grad_norm": 3.7875027656555176, | |
| "learning_rate": 9.195402298850575e-05, | |
| "loss": 1.3364, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 6.068965517241379, | |
| "grad_norm": 3.290332078933716, | |
| "learning_rate": 9.184455391351943e-05, | |
| "loss": 1.4505, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 6.103448275862069, | |
| "grad_norm": 2.82051682472229, | |
| "learning_rate": 9.173508483853312e-05, | |
| "loss": 1.2286, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 6.137931034482759, | |
| "grad_norm": 2.4758310317993164, | |
| "learning_rate": 9.16256157635468e-05, | |
| "loss": 1.6257, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 6.172413793103448, | |
| "grad_norm": 2.9103426933288574, | |
| "learning_rate": 9.151614668856049e-05, | |
| "loss": 1.2839, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 6.206896551724138, | |
| "grad_norm": 3.811338424682617, | |
| "learning_rate": 9.140667761357416e-05, | |
| "loss": 1.5439, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 6.241379310344827, | |
| "grad_norm": 2.94343638420105, | |
| "learning_rate": 9.129720853858785e-05, | |
| "loss": 1.6658, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 6.275862068965517, | |
| "grad_norm": 3.5924301147460938, | |
| "learning_rate": 9.118773946360154e-05, | |
| "loss": 1.3065, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 6.310344827586207, | |
| "grad_norm": 2.893584728240967, | |
| "learning_rate": 9.107827038861522e-05, | |
| "loss": 1.7638, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 6.344827586206897, | |
| "grad_norm": 4.023632049560547, | |
| "learning_rate": 9.096880131362891e-05, | |
| "loss": 1.1021, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 6.379310344827586, | |
| "grad_norm": 2.928293228149414, | |
| "learning_rate": 9.08593322386426e-05, | |
| "loss": 1.3904, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 6.413793103448276, | |
| "grad_norm": 3.4317667484283447, | |
| "learning_rate": 9.074986316365627e-05, | |
| "loss": 1.3181, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 6.448275862068965, | |
| "grad_norm": 2.3280980587005615, | |
| "learning_rate": 9.064039408866995e-05, | |
| "loss": 1.0595, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 6.482758620689655, | |
| "grad_norm": 2.5254228115081787, | |
| "learning_rate": 9.053092501368364e-05, | |
| "loss": 1.2124, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 6.517241379310345, | |
| "grad_norm": 2.9616031646728516, | |
| "learning_rate": 9.042145593869731e-05, | |
| "loss": 1.4659, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 6.551724137931035, | |
| "grad_norm": 3.044532299041748, | |
| "learning_rate": 9.031198686371101e-05, | |
| "loss": 1.5561, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 6.586206896551724, | |
| "grad_norm": 4.845521450042725, | |
| "learning_rate": 9.02025177887247e-05, | |
| "loss": 1.3714, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 6.620689655172414, | |
| "grad_norm": 2.6942129135131836, | |
| "learning_rate": 9.009304871373837e-05, | |
| "loss": 1.0138, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 6.655172413793103, | |
| "grad_norm": 3.31146502494812, | |
| "learning_rate": 8.998357963875206e-05, | |
| "loss": 1.2017, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 6.689655172413794, | |
| "grad_norm": 4.051918029785156, | |
| "learning_rate": 8.987411056376574e-05, | |
| "loss": 1.6548, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 6.724137931034483, | |
| "grad_norm": 2.59892201423645, | |
| "learning_rate": 8.976464148877942e-05, | |
| "loss": 1.6622, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 6.758620689655173, | |
| "grad_norm": 3.7553224563598633, | |
| "learning_rate": 8.96551724137931e-05, | |
| "loss": 1.7852, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 6.793103448275862, | |
| "grad_norm": 3.6946635246276855, | |
| "learning_rate": 8.95457033388068e-05, | |
| "loss": 1.1174, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 6.827586206896552, | |
| "grad_norm": 4.1242194175720215, | |
| "learning_rate": 8.943623426382048e-05, | |
| "loss": 1.4384, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 6.862068965517241, | |
| "grad_norm": 2.6595194339752197, | |
| "learning_rate": 8.932676518883416e-05, | |
| "loss": 1.2971, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 6.896551724137931, | |
| "grad_norm": 9.484551429748535, | |
| "learning_rate": 8.921729611384783e-05, | |
| "loss": 1.1471, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 6.931034482758621, | |
| "grad_norm": 6.137091636657715, | |
| "learning_rate": 8.910782703886152e-05, | |
| "loss": 1.5765, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 6.9655172413793105, | |
| "grad_norm": 3.7353861331939697, | |
| "learning_rate": 8.899835796387521e-05, | |
| "loss": 1.4145, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "grad_norm": 4.387484073638916, | |
| "learning_rate": 8.888888888888889e-05, | |
| "loss": 1.4543, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_accuracy": 0.5151515151515151, | |
| "eval_f1_macro": 0.4112892443400918, | |
| "eval_f1_micro": 0.5151515151515151, | |
| "eval_f1_weighted": 0.4707506497953339, | |
| "eval_loss": 1.3674346208572388, | |
| "eval_precision_macro": 0.4076719576719577, | |
| "eval_precision_micro": 0.5151515151515151, | |
| "eval_precision_weighted": 0.46301247771836007, | |
| "eval_recall_macro": 0.44794406651549507, | |
| "eval_recall_micro": 0.5151515151515151, | |
| "eval_recall_weighted": 0.5151515151515151, | |
| "eval_runtime": 1.681, | |
| "eval_samples_per_second": 78.524, | |
| "eval_steps_per_second": 10.113, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 7.0344827586206895, | |
| "grad_norm": 5.052458763122559, | |
| "learning_rate": 8.877941981390258e-05, | |
| "loss": 1.1215, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 7.068965517241379, | |
| "grad_norm": 2.47261905670166, | |
| "learning_rate": 8.866995073891627e-05, | |
| "loss": 1.0094, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 7.103448275862069, | |
| "grad_norm": 3.871811866760254, | |
| "learning_rate": 8.856048166392994e-05, | |
| "loss": 1.3071, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 7.137931034482759, | |
| "grad_norm": 4.654438018798828, | |
| "learning_rate": 8.845101258894362e-05, | |
| "loss": 1.5312, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 7.172413793103448, | |
| "grad_norm": 5.699710845947266, | |
| "learning_rate": 8.834154351395731e-05, | |
| "loss": 1.376, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 7.206896551724138, | |
| "grad_norm": 5.32385778427124, | |
| "learning_rate": 8.8232074438971e-05, | |
| "loss": 1.3338, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 7.241379310344827, | |
| "grad_norm": 2.77158522605896, | |
| "learning_rate": 8.812260536398468e-05, | |
| "loss": 1.2957, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 7.275862068965517, | |
| "grad_norm": 2.6451570987701416, | |
| "learning_rate": 8.801313628899837e-05, | |
| "loss": 1.5218, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 7.310344827586207, | |
| "grad_norm": 5.987870216369629, | |
| "learning_rate": 8.790366721401204e-05, | |
| "loss": 1.5623, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 7.344827586206897, | |
| "grad_norm": 2.895263195037842, | |
| "learning_rate": 8.779419813902573e-05, | |
| "loss": 1.1866, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 7.379310344827586, | |
| "grad_norm": 5.6424360275268555, | |
| "learning_rate": 8.768472906403941e-05, | |
| "loss": 0.9502, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 7.413793103448276, | |
| "grad_norm": 9.878355026245117, | |
| "learning_rate": 8.757525998905309e-05, | |
| "loss": 1.5713, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 7.448275862068965, | |
| "grad_norm": 3.341871738433838, | |
| "learning_rate": 8.746579091406679e-05, | |
| "loss": 1.5982, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 7.482758620689655, | |
| "grad_norm": 5.839073181152344, | |
| "learning_rate": 8.735632183908047e-05, | |
| "loss": 1.6279, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 7.517241379310345, | |
| "grad_norm": 3.4865572452545166, | |
| "learning_rate": 8.724685276409415e-05, | |
| "loss": 1.5146, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 7.551724137931035, | |
| "grad_norm": 3.3559305667877197, | |
| "learning_rate": 8.713738368910783e-05, | |
| "loss": 1.2828, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 7.586206896551724, | |
| "grad_norm": 5.358764171600342, | |
| "learning_rate": 8.702791461412152e-05, | |
| "loss": 1.4213, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 7.620689655172414, | |
| "grad_norm": 3.7849018573760986, | |
| "learning_rate": 8.691844553913519e-05, | |
| "loss": 1.1776, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 7.655172413793103, | |
| "grad_norm": 5.408728122711182, | |
| "learning_rate": 8.680897646414888e-05, | |
| "loss": 1.2171, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 7.689655172413794, | |
| "grad_norm": 3.210322618484497, | |
| "learning_rate": 8.669950738916258e-05, | |
| "loss": 1.139, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 7.724137931034483, | |
| "grad_norm": 4.863003730773926, | |
| "learning_rate": 8.659003831417625e-05, | |
| "loss": 1.0228, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 7.758620689655173, | |
| "grad_norm": 2.4240548610687256, | |
| "learning_rate": 8.648056923918994e-05, | |
| "loss": 1.1159, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 7.793103448275862, | |
| "grad_norm": 4.020617961883545, | |
| "learning_rate": 8.637110016420362e-05, | |
| "loss": 1.6475, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 7.827586206896552, | |
| "grad_norm": 3.7977688312530518, | |
| "learning_rate": 8.62616310892173e-05, | |
| "loss": 1.0921, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 7.862068965517241, | |
| "grad_norm": 6.774123191833496, | |
| "learning_rate": 8.615216201423098e-05, | |
| "loss": 1.1647, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 7.896551724137931, | |
| "grad_norm": 3.9084153175354004, | |
| "learning_rate": 8.604269293924467e-05, | |
| "loss": 1.5981, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 7.931034482758621, | |
| "grad_norm": 3.0181827545166016, | |
| "learning_rate": 8.593322386425835e-05, | |
| "loss": 1.4641, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 7.9655172413793105, | |
| "grad_norm": 5.793522357940674, | |
| "learning_rate": 8.582375478927204e-05, | |
| "loss": 1.4984, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 3.726046085357666, | |
| "learning_rate": 8.571428571428571e-05, | |
| "loss": 1.2349, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_accuracy": 0.5, | |
| "eval_f1_macro": 0.40237683828026144, | |
| "eval_f1_micro": 0.5, | |
| "eval_f1_weighted": 0.45497924943822293, | |
| "eval_loss": 1.313656210899353, | |
| "eval_precision_macro": 0.40499231950844855, | |
| "eval_precision_micro": 0.5, | |
| "eval_precision_weighted": 0.46061013359400454, | |
| "eval_recall_macro": 0.4478835978835979, | |
| "eval_recall_micro": 0.5, | |
| "eval_recall_weighted": 0.5, | |
| "eval_runtime": 1.7557, | |
| "eval_samples_per_second": 75.184, | |
| "eval_steps_per_second": 9.683, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 8.03448275862069, | |
| "grad_norm": 1.9977046251296997, | |
| "learning_rate": 8.56048166392994e-05, | |
| "loss": 1.0337, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 8.068965517241379, | |
| "grad_norm": 3.49200701713562, | |
| "learning_rate": 8.549534756431309e-05, | |
| "loss": 1.1619, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 8.10344827586207, | |
| "grad_norm": 6.4927825927734375, | |
| "learning_rate": 8.538587848932677e-05, | |
| "loss": 1.4676, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 8.137931034482758, | |
| "grad_norm": 1.9825376272201538, | |
| "learning_rate": 8.527640941434046e-05, | |
| "loss": 1.0621, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 8.172413793103448, | |
| "grad_norm": 3.2611658573150635, | |
| "learning_rate": 8.516694033935414e-05, | |
| "loss": 1.4685, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 8.206896551724139, | |
| "grad_norm": 6.502848148345947, | |
| "learning_rate": 8.505747126436782e-05, | |
| "loss": 1.7512, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 8.241379310344827, | |
| "grad_norm": 3.67802357673645, | |
| "learning_rate": 8.49480021893815e-05, | |
| "loss": 1.2843, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 8.275862068965518, | |
| "grad_norm": 3.7158236503601074, | |
| "learning_rate": 8.483853311439519e-05, | |
| "loss": 1.2995, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 8.310344827586206, | |
| "grad_norm": 2.446770191192627, | |
| "learning_rate": 8.472906403940886e-05, | |
| "loss": 1.2817, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 8.344827586206897, | |
| "grad_norm": 3.7978079319000244, | |
| "learning_rate": 8.461959496442256e-05, | |
| "loss": 1.1886, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 8.379310344827585, | |
| "grad_norm": 4.225602149963379, | |
| "learning_rate": 8.451012588943625e-05, | |
| "loss": 1.1122, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 8.413793103448276, | |
| "grad_norm": 4.149050712585449, | |
| "learning_rate": 8.440065681444992e-05, | |
| "loss": 1.2523, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 8.448275862068966, | |
| "grad_norm": 1.8350954055786133, | |
| "learning_rate": 8.42911877394636e-05, | |
| "loss": 1.2488, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 8.482758620689655, | |
| "grad_norm": 5.763144493103027, | |
| "learning_rate": 8.418171866447729e-05, | |
| "loss": 1.2622, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 8.517241379310345, | |
| "grad_norm": 3.465343952178955, | |
| "learning_rate": 8.407224958949097e-05, | |
| "loss": 1.1619, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 8.551724137931034, | |
| "grad_norm": 3.9785590171813965, | |
| "learning_rate": 8.396278051450465e-05, | |
| "loss": 1.3165, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 8.586206896551724, | |
| "grad_norm": 3.5696113109588623, | |
| "learning_rate": 8.385331143951835e-05, | |
| "loss": 1.2519, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 8.620689655172415, | |
| "grad_norm": 7.4920878410339355, | |
| "learning_rate": 8.374384236453202e-05, | |
| "loss": 1.1137, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 8.655172413793103, | |
| "grad_norm": 3.3087165355682373, | |
| "learning_rate": 8.363437328954571e-05, | |
| "loss": 1.0042, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 8.689655172413794, | |
| "grad_norm": 5.924715518951416, | |
| "learning_rate": 8.35249042145594e-05, | |
| "loss": 1.5938, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 8.724137931034482, | |
| "grad_norm": 4.854137897491455, | |
| "learning_rate": 8.341543513957307e-05, | |
| "loss": 1.6971, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 8.758620689655173, | |
| "grad_norm": 2.877127170562744, | |
| "learning_rate": 8.330596606458676e-05, | |
| "loss": 1.4602, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 8.793103448275861, | |
| "grad_norm": 4.012562274932861, | |
| "learning_rate": 8.319649698960044e-05, | |
| "loss": 1.1416, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 8.827586206896552, | |
| "grad_norm": 7.408483028411865, | |
| "learning_rate": 8.308702791461413e-05, | |
| "loss": 1.3352, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 8.862068965517242, | |
| "grad_norm": 3.3066165447235107, | |
| "learning_rate": 8.297755883962781e-05, | |
| "loss": 1.3075, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 8.89655172413793, | |
| "grad_norm": 5.024322509765625, | |
| "learning_rate": 8.286808976464149e-05, | |
| "loss": 1.3557, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 8.931034482758621, | |
| "grad_norm": 4.212279319763184, | |
| "learning_rate": 8.275862068965517e-05, | |
| "loss": 1.103, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 8.96551724137931, | |
| "grad_norm": 4.724401473999023, | |
| "learning_rate": 8.264915161466886e-05, | |
| "loss": 0.9081, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "grad_norm": 4.074641704559326, | |
| "learning_rate": 8.253968253968255e-05, | |
| "loss": 1.2544, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_accuracy": 0.5075757575757576, | |
| "eval_f1_macro": 0.42087451332734344, | |
| "eval_f1_micro": 0.5075757575757576, | |
| "eval_f1_weighted": 0.47479591512610375, | |
| "eval_loss": 1.3322217464447021, | |
| "eval_precision_macro": 0.4223733938019652, | |
| "eval_precision_micro": 0.5075757575757576, | |
| "eval_precision_weighted": 0.4737253487253487, | |
| "eval_recall_macro": 0.447989417989418, | |
| "eval_recall_micro": 0.5075757575757576, | |
| "eval_recall_weighted": 0.5075757575757576, | |
| "eval_runtime": 1.7338, | |
| "eval_samples_per_second": 76.133, | |
| "eval_steps_per_second": 9.805, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 9.03448275862069, | |
| "grad_norm": 3.3098955154418945, | |
| "learning_rate": 8.243021346469623e-05, | |
| "loss": 0.9961, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 9.068965517241379, | |
| "grad_norm": 4.669078826904297, | |
| "learning_rate": 8.232074438970992e-05, | |
| "loss": 0.8428, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 9.10344827586207, | |
| "grad_norm": 10.585116386413574, | |
| "learning_rate": 8.221127531472359e-05, | |
| "loss": 1.093, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 9.137931034482758, | |
| "grad_norm": 3.542888641357422, | |
| "learning_rate": 8.210180623973728e-05, | |
| "loss": 1.1071, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 9.172413793103448, | |
| "grad_norm": 3.5215578079223633, | |
| "learning_rate": 8.199233716475096e-05, | |
| "loss": 1.3982, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 9.206896551724139, | |
| "grad_norm": 2.9111316204071045, | |
| "learning_rate": 8.188286808976464e-05, | |
| "loss": 1.3159, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 9.241379310344827, | |
| "grad_norm": 3.4728190898895264, | |
| "learning_rate": 8.177339901477834e-05, | |
| "loss": 0.8367, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 9.275862068965518, | |
| "grad_norm": 7.045929908752441, | |
| "learning_rate": 8.166392993979202e-05, | |
| "loss": 1.1379, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 9.310344827586206, | |
| "grad_norm": 3.804302215576172, | |
| "learning_rate": 8.15544608648057e-05, | |
| "loss": 1.7455, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 9.344827586206897, | |
| "grad_norm": 5.53285551071167, | |
| "learning_rate": 8.144499178981938e-05, | |
| "loss": 0.8734, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 9.379310344827585, | |
| "grad_norm": 4.002763271331787, | |
| "learning_rate": 8.133552271483307e-05, | |
| "loss": 1.2031, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 9.413793103448276, | |
| "grad_norm": 3.3367197513580322, | |
| "learning_rate": 8.122605363984674e-05, | |
| "loss": 1.3156, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 9.448275862068966, | |
| "grad_norm": 6.759856700897217, | |
| "learning_rate": 8.111658456486043e-05, | |
| "loss": 1.5352, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 9.482758620689655, | |
| "grad_norm": 3.184265375137329, | |
| "learning_rate": 8.100711548987413e-05, | |
| "loss": 0.8941, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 9.517241379310345, | |
| "grad_norm": 5.55509090423584, | |
| "learning_rate": 8.08976464148878e-05, | |
| "loss": 1.1496, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 9.551724137931034, | |
| "grad_norm": 3.5079853534698486, | |
| "learning_rate": 8.078817733990148e-05, | |
| "loss": 1.0693, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 9.586206896551724, | |
| "grad_norm": 3.0940561294555664, | |
| "learning_rate": 8.067870826491517e-05, | |
| "loss": 1.0685, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 9.620689655172415, | |
| "grad_norm": 3.501739978790283, | |
| "learning_rate": 8.056923918992884e-05, | |
| "loss": 1.0207, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 9.655172413793103, | |
| "grad_norm": 2.481731414794922, | |
| "learning_rate": 8.045977011494253e-05, | |
| "loss": 0.7245, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 9.689655172413794, | |
| "grad_norm": 3.4049670696258545, | |
| "learning_rate": 8.035030103995622e-05, | |
| "loss": 1.4967, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 9.724137931034482, | |
| "grad_norm": 2.143556833267212, | |
| "learning_rate": 8.02408319649699e-05, | |
| "loss": 1.3994, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 9.758620689655173, | |
| "grad_norm": 3.563452959060669, | |
| "learning_rate": 8.013136288998359e-05, | |
| "loss": 1.3212, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 9.793103448275861, | |
| "grad_norm": 3.2039382457733154, | |
| "learning_rate": 8.002189381499726e-05, | |
| "loss": 0.9787, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 9.827586206896552, | |
| "grad_norm": 3.1124753952026367, | |
| "learning_rate": 7.991242474001095e-05, | |
| "loss": 0.9052, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 9.862068965517242, | |
| "grad_norm": 4.318131923675537, | |
| "learning_rate": 7.980295566502463e-05, | |
| "loss": 1.4156, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 9.89655172413793, | |
| "grad_norm": 3.689425230026245, | |
| "learning_rate": 7.969348659003832e-05, | |
| "loss": 1.1676, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 9.931034482758621, | |
| "grad_norm": 4.3956499099731445, | |
| "learning_rate": 7.9584017515052e-05, | |
| "loss": 1.1724, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 9.96551724137931, | |
| "grad_norm": 4.972238540649414, | |
| "learning_rate": 7.947454844006569e-05, | |
| "loss": 1.5352, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 4.761918544769287, | |
| "learning_rate": 7.936507936507937e-05, | |
| "loss": 1.206, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_accuracy": 0.4621212121212121, | |
| "eval_f1_macro": 0.35546523346854747, | |
| "eval_f1_micro": 0.4621212121212121, | |
| "eval_f1_weighted": 0.4008642480388733, | |
| "eval_loss": 1.3817532062530518, | |
| "eval_precision_macro": 0.3931256296985198, | |
| "eval_precision_micro": 0.4621212121212121, | |
| "eval_precision_weighted": 0.43718074725398604, | |
| "eval_recall_macro": 0.4129402872260015, | |
| "eval_recall_micro": 0.4621212121212121, | |
| "eval_recall_weighted": 0.4621212121212121, | |
| "eval_runtime": 1.703, | |
| "eval_samples_per_second": 77.512, | |
| "eval_steps_per_second": 9.983, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 10.03448275862069, | |
| "grad_norm": 3.7589895725250244, | |
| "learning_rate": 7.925561029009305e-05, | |
| "loss": 1.096, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 10.068965517241379, | |
| "grad_norm": 5.561600685119629, | |
| "learning_rate": 7.914614121510674e-05, | |
| "loss": 1.381, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 10.10344827586207, | |
| "grad_norm": 4.066978454589844, | |
| "learning_rate": 7.903667214012041e-05, | |
| "loss": 1.1099, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 10.137931034482758, | |
| "grad_norm": 2.7694287300109863, | |
| "learning_rate": 7.892720306513411e-05, | |
| "loss": 0.8646, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 10.172413793103448, | |
| "grad_norm": 3.7669525146484375, | |
| "learning_rate": 7.88177339901478e-05, | |
| "loss": 1.02, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 10.206896551724139, | |
| "grad_norm": 13.383344650268555, | |
| "learning_rate": 7.870826491516147e-05, | |
| "loss": 1.7312, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 10.241379310344827, | |
| "grad_norm": 2.007171154022217, | |
| "learning_rate": 7.859879584017516e-05, | |
| "loss": 1.0665, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 10.275862068965518, | |
| "grad_norm": 4.019918441772461, | |
| "learning_rate": 7.848932676518884e-05, | |
| "loss": 1.2016, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 10.310344827586206, | |
| "grad_norm": 8.339564323425293, | |
| "learning_rate": 7.837985769020251e-05, | |
| "loss": 1.7457, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 10.344827586206897, | |
| "grad_norm": 2.7739768028259277, | |
| "learning_rate": 7.82703886152162e-05, | |
| "loss": 1.3515, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 10.379310344827585, | |
| "grad_norm": 5.876687526702881, | |
| "learning_rate": 7.81609195402299e-05, | |
| "loss": 0.9722, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 10.413793103448276, | |
| "grad_norm": 3.3724043369293213, | |
| "learning_rate": 7.805145046524357e-05, | |
| "loss": 0.9159, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 10.448275862068966, | |
| "grad_norm": 2.4745233058929443, | |
| "learning_rate": 7.794198139025726e-05, | |
| "loss": 1.0241, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 10.482758620689655, | |
| "grad_norm": 6.936390399932861, | |
| "learning_rate": 7.783251231527095e-05, | |
| "loss": 1.11, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 10.517241379310345, | |
| "grad_norm": 3.721477746963501, | |
| "learning_rate": 7.772304324028462e-05, | |
| "loss": 1.396, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 10.551724137931034, | |
| "grad_norm": 3.1573660373687744, | |
| "learning_rate": 7.76135741652983e-05, | |
| "loss": 0.925, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 10.586206896551724, | |
| "grad_norm": 3.663121461868286, | |
| "learning_rate": 7.750410509031199e-05, | |
| "loss": 1.0975, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 10.620689655172415, | |
| "grad_norm": 2.9894790649414062, | |
| "learning_rate": 7.739463601532568e-05, | |
| "loss": 1.3367, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 10.655172413793103, | |
| "grad_norm": 4.422184467315674, | |
| "learning_rate": 7.728516694033936e-05, | |
| "loss": 0.8371, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 10.689655172413794, | |
| "grad_norm": 3.6148533821105957, | |
| "learning_rate": 7.717569786535304e-05, | |
| "loss": 1.6326, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 10.724137931034482, | |
| "grad_norm": 2.576557159423828, | |
| "learning_rate": 7.706622879036672e-05, | |
| "loss": 1.0553, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 10.758620689655173, | |
| "grad_norm": 3.054694652557373, | |
| "learning_rate": 7.695675971538041e-05, | |
| "loss": 0.9281, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 10.793103448275861, | |
| "grad_norm": 5.310608863830566, | |
| "learning_rate": 7.684729064039408e-05, | |
| "loss": 1.1227, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 10.827586206896552, | |
| "grad_norm": 10.316115379333496, | |
| "learning_rate": 7.673782156540778e-05, | |
| "loss": 0.8636, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 10.862068965517242, | |
| "grad_norm": 2.8177926540374756, | |
| "learning_rate": 7.662835249042147e-05, | |
| "loss": 0.6684, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 10.89655172413793, | |
| "grad_norm": 8.028295516967773, | |
| "learning_rate": 7.651888341543514e-05, | |
| "loss": 0.6878, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 10.931034482758621, | |
| "grad_norm": 2.236884355545044, | |
| "learning_rate": 7.640941434044883e-05, | |
| "loss": 0.8953, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 10.96551724137931, | |
| "grad_norm": 3.125170946121216, | |
| "learning_rate": 7.629994526546251e-05, | |
| "loss": 0.7141, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "grad_norm": 3.6889586448669434, | |
| "learning_rate": 7.619047619047618e-05, | |
| "loss": 1.0416, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_accuracy": 0.5606060606060606, | |
| "eval_f1_macro": 0.46100235596487465, | |
| "eval_f1_micro": 0.5606060606060606, | |
| "eval_f1_weighted": 0.5249136795238745, | |
| "eval_loss": 1.314196228981018, | |
| "eval_precision_macro": 0.5218135772974483, | |
| "eval_precision_micro": 0.5606060606060606, | |
| "eval_precision_weighted": 0.5872401994102874, | |
| "eval_recall_macro": 0.49511715797430084, | |
| "eval_recall_micro": 0.5606060606060606, | |
| "eval_recall_weighted": 0.5606060606060606, | |
| "eval_runtime": 1.7117, | |
| "eval_samples_per_second": 77.117, | |
| "eval_steps_per_second": 9.932, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 11.03448275862069, | |
| "grad_norm": 10.25067138671875, | |
| "learning_rate": 7.608100711548987e-05, | |
| "loss": 1.0965, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 11.068965517241379, | |
| "grad_norm": 2.579963207244873, | |
| "learning_rate": 7.597153804050357e-05, | |
| "loss": 1.1408, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 11.10344827586207, | |
| "grad_norm": 3.9004721641540527, | |
| "learning_rate": 7.586206896551724e-05, | |
| "loss": 1.1872, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 11.137931034482758, | |
| "grad_norm": 3.8652117252349854, | |
| "learning_rate": 7.575259989053093e-05, | |
| "loss": 0.8976, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 11.172413793103448, | |
| "grad_norm": 4.144958972930908, | |
| "learning_rate": 7.564313081554462e-05, | |
| "loss": 0.7235, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 11.206896551724139, | |
| "grad_norm": 2.5177595615386963, | |
| "learning_rate": 7.553366174055829e-05, | |
| "loss": 0.8665, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 11.241379310344827, | |
| "grad_norm": 4.09409236907959, | |
| "learning_rate": 7.542419266557197e-05, | |
| "loss": 0.7202, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 11.275862068965518, | |
| "grad_norm": 6.234818458557129, | |
| "learning_rate": 7.531472359058566e-05, | |
| "loss": 0.9528, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 11.310344827586206, | |
| "grad_norm": 6.615692615509033, | |
| "learning_rate": 7.520525451559935e-05, | |
| "loss": 1.1562, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 11.344827586206897, | |
| "grad_norm": 3.72806978225708, | |
| "learning_rate": 7.509578544061303e-05, | |
| "loss": 1.2677, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 11.379310344827585, | |
| "grad_norm": 4.646265506744385, | |
| "learning_rate": 7.498631636562672e-05, | |
| "loss": 1.3499, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 11.413793103448276, | |
| "grad_norm": 2.4671194553375244, | |
| "learning_rate": 7.487684729064039e-05, | |
| "loss": 0.9052, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 11.448275862068966, | |
| "grad_norm": 3.04329514503479, | |
| "learning_rate": 7.476737821565408e-05, | |
| "loss": 1.2754, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 11.482758620689655, | |
| "grad_norm": 2.1641273498535156, | |
| "learning_rate": 7.465790914066776e-05, | |
| "loss": 0.9117, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 11.517241379310345, | |
| "grad_norm": 2.3654139041900635, | |
| "learning_rate": 7.454844006568145e-05, | |
| "loss": 0.9465, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 11.551724137931034, | |
| "grad_norm": 9.097860336303711, | |
| "learning_rate": 7.443897099069514e-05, | |
| "loss": 1.2733, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 11.586206896551724, | |
| "grad_norm": 4.378310680389404, | |
| "learning_rate": 7.432950191570882e-05, | |
| "loss": 0.9539, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 11.620689655172415, | |
| "grad_norm": 5.611904621124268, | |
| "learning_rate": 7.42200328407225e-05, | |
| "loss": 1.056, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 11.655172413793103, | |
| "grad_norm": 3.2342333793640137, | |
| "learning_rate": 7.411056376573618e-05, | |
| "loss": 0.9286, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 11.689655172413794, | |
| "grad_norm": 4.163808822631836, | |
| "learning_rate": 7.400109469074985e-05, | |
| "loss": 0.8409, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 11.724137931034482, | |
| "grad_norm": 3.296419382095337, | |
| "learning_rate": 7.389162561576355e-05, | |
| "loss": 0.9829, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 11.758620689655173, | |
| "grad_norm": 6.1151227951049805, | |
| "learning_rate": 7.378215654077724e-05, | |
| "loss": 1.2809, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 11.793103448275861, | |
| "grad_norm": 6.986661911010742, | |
| "learning_rate": 7.367268746579091e-05, | |
| "loss": 1.0579, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 11.827586206896552, | |
| "grad_norm": 3.160599946975708, | |
| "learning_rate": 7.35632183908046e-05, | |
| "loss": 0.805, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 11.862068965517242, | |
| "grad_norm": 2.7907633781433105, | |
| "learning_rate": 7.345374931581829e-05, | |
| "loss": 0.8927, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 11.89655172413793, | |
| "grad_norm": 6.923659324645996, | |
| "learning_rate": 7.334428024083196e-05, | |
| "loss": 1.1211, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 11.931034482758621, | |
| "grad_norm": 5.215874195098877, | |
| "learning_rate": 7.323481116584564e-05, | |
| "loss": 1.1497, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 11.96551724137931, | |
| "grad_norm": 3.4313321113586426, | |
| "learning_rate": 7.312534209085934e-05, | |
| "loss": 0.8678, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "grad_norm": 4.915594100952148, | |
| "learning_rate": 7.301587301587302e-05, | |
| "loss": 1.1494, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_accuracy": 0.4772727272727273, | |
| "eval_f1_macro": 0.41057591774355634, | |
| "eval_f1_micro": 0.4772727272727273, | |
| "eval_f1_weighted": 0.4652476264825862, | |
| "eval_loss": 1.3793162107467651, | |
| "eval_precision_macro": 0.46187641723356004, | |
| "eval_precision_micro": 0.4772727272727273, | |
| "eval_precision_weighted": 0.5255907287157288, | |
| "eval_recall_macro": 0.4227286470143613, | |
| "eval_recall_micro": 0.4772727272727273, | |
| "eval_recall_weighted": 0.4772727272727273, | |
| "eval_runtime": 1.691, | |
| "eval_samples_per_second": 78.06, | |
| "eval_steps_per_second": 10.053, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 12.03448275862069, | |
| "grad_norm": 6.37459659576416, | |
| "learning_rate": 7.29064039408867e-05, | |
| "loss": 1.2047, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 12.068965517241379, | |
| "grad_norm": 4.954467296600342, | |
| "learning_rate": 7.279693486590039e-05, | |
| "loss": 1.0239, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 12.10344827586207, | |
| "grad_norm": 2.187420129776001, | |
| "learning_rate": 7.268746579091406e-05, | |
| "loss": 0.5644, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 12.137931034482758, | |
| "grad_norm": 4.12716817855835, | |
| "learning_rate": 7.257799671592775e-05, | |
| "loss": 0.7901, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 12.172413793103448, | |
| "grad_norm": 4.0836873054504395, | |
| "learning_rate": 7.246852764094144e-05, | |
| "loss": 1.3414, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 12.206896551724139, | |
| "grad_norm": 4.051703929901123, | |
| "learning_rate": 7.235905856595512e-05, | |
| "loss": 1.1838, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 12.241379310344827, | |
| "grad_norm": 2.682950973510742, | |
| "learning_rate": 7.224958949096881e-05, | |
| "loss": 0.7483, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 12.275862068965518, | |
| "grad_norm": 2.350590944290161, | |
| "learning_rate": 7.21401204159825e-05, | |
| "loss": 0.6146, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 12.310344827586206, | |
| "grad_norm": 4.472531795501709, | |
| "learning_rate": 7.203065134099617e-05, | |
| "loss": 1.0984, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 12.344827586206897, | |
| "grad_norm": 3.710747480392456, | |
| "learning_rate": 7.192118226600985e-05, | |
| "loss": 0.7819, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 12.379310344827585, | |
| "grad_norm": 6.625567436218262, | |
| "learning_rate": 7.181171319102354e-05, | |
| "loss": 0.9725, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 12.413793103448276, | |
| "grad_norm": 5.959197998046875, | |
| "learning_rate": 7.170224411603723e-05, | |
| "loss": 1.0577, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 12.448275862068966, | |
| "grad_norm": 3.8905091285705566, | |
| "learning_rate": 7.159277504105091e-05, | |
| "loss": 0.7902, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 12.482758620689655, | |
| "grad_norm": 9.632880210876465, | |
| "learning_rate": 7.14833059660646e-05, | |
| "loss": 0.8034, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 12.517241379310345, | |
| "grad_norm": 2.0040698051452637, | |
| "learning_rate": 7.137383689107827e-05, | |
| "loss": 0.6375, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 12.551724137931034, | |
| "grad_norm": 3.397597074508667, | |
| "learning_rate": 7.126436781609196e-05, | |
| "loss": 0.9355, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 12.586206896551724, | |
| "grad_norm": 4.660778522491455, | |
| "learning_rate": 7.115489874110563e-05, | |
| "loss": 0.97, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 12.620689655172415, | |
| "grad_norm": 4.888659954071045, | |
| "learning_rate": 7.104542966611933e-05, | |
| "loss": 1.0236, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 12.655172413793103, | |
| "grad_norm": 2.9293081760406494, | |
| "learning_rate": 7.093596059113302e-05, | |
| "loss": 0.8456, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 12.689655172413794, | |
| "grad_norm": 7.08192253112793, | |
| "learning_rate": 7.082649151614669e-05, | |
| "loss": 1.3017, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 12.724137931034482, | |
| "grad_norm": 4.688408374786377, | |
| "learning_rate": 7.071702244116037e-05, | |
| "loss": 1.3321, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 12.758620689655173, | |
| "grad_norm": 6.150996685028076, | |
| "learning_rate": 7.060755336617406e-05, | |
| "loss": 1.5606, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 12.793103448275861, | |
| "grad_norm": 6.134938716888428, | |
| "learning_rate": 7.049808429118773e-05, | |
| "loss": 1.0207, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 12.827586206896552, | |
| "grad_norm": 4.253851890563965, | |
| "learning_rate": 7.038861521620142e-05, | |
| "loss": 0.637, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 12.862068965517242, | |
| "grad_norm": 2.356126546859741, | |
| "learning_rate": 7.027914614121512e-05, | |
| "loss": 0.5845, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 12.89655172413793, | |
| "grad_norm": 6.397014141082764, | |
| "learning_rate": 7.016967706622879e-05, | |
| "loss": 1.1618, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 12.931034482758621, | |
| "grad_norm": 2.425175666809082, | |
| "learning_rate": 7.006020799124248e-05, | |
| "loss": 0.8193, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 12.96551724137931, | |
| "grad_norm": 4.115069389343262, | |
| "learning_rate": 6.995073891625616e-05, | |
| "loss": 1.0236, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "grad_norm": 3.485793113708496, | |
| "learning_rate": 6.984126984126984e-05, | |
| "loss": 0.7366, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5655740899926947, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6382644888194571, | |
| "eval_loss": 1.1935979127883911, | |
| "eval_precision_macro": 0.5707684918779371, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6445614261104243, | |
| "eval_recall_macro": 0.5790249433106576, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 1.6957, | |
| "eval_samples_per_second": 77.845, | |
| "eval_steps_per_second": 10.026, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 13.03448275862069, | |
| "grad_norm": 3.7799136638641357, | |
| "learning_rate": 6.973180076628352e-05, | |
| "loss": 0.5383, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 13.068965517241379, | |
| "grad_norm": 4.496936798095703, | |
| "learning_rate": 6.962233169129721e-05, | |
| "loss": 0.9326, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 13.10344827586207, | |
| "grad_norm": 2.7297375202178955, | |
| "learning_rate": 6.95128626163109e-05, | |
| "loss": 0.6216, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 13.137931034482758, | |
| "grad_norm": 4.5329179763793945, | |
| "learning_rate": 6.940339354132458e-05, | |
| "loss": 1.0146, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 13.172413793103448, | |
| "grad_norm": 4.96497917175293, | |
| "learning_rate": 6.929392446633827e-05, | |
| "loss": 0.8865, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 13.206896551724139, | |
| "grad_norm": 3.1082141399383545, | |
| "learning_rate": 6.918445539135194e-05, | |
| "loss": 0.8833, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 13.241379310344827, | |
| "grad_norm": 6.976032733917236, | |
| "learning_rate": 6.907498631636563e-05, | |
| "loss": 0.7684, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 13.275862068965518, | |
| "grad_norm": 1.962904691696167, | |
| "learning_rate": 6.896551724137931e-05, | |
| "loss": 0.3621, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 13.310344827586206, | |
| "grad_norm": 3.1863770484924316, | |
| "learning_rate": 6.8856048166393e-05, | |
| "loss": 0.3451, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 13.344827586206897, | |
| "grad_norm": 2.953831672668457, | |
| "learning_rate": 6.874657909140669e-05, | |
| "loss": 0.6924, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 13.379310344827585, | |
| "grad_norm": 4.513678073883057, | |
| "learning_rate": 6.863711001642037e-05, | |
| "loss": 0.6521, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 13.413793103448276, | |
| "grad_norm": 2.1181821823120117, | |
| "learning_rate": 6.852764094143404e-05, | |
| "loss": 0.7305, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 13.448275862068966, | |
| "grad_norm": 5.158665657043457, | |
| "learning_rate": 6.841817186644773e-05, | |
| "loss": 0.757, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 13.482758620689655, | |
| "grad_norm": 2.678704261779785, | |
| "learning_rate": 6.830870279146142e-05, | |
| "loss": 0.7164, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 13.517241379310345, | |
| "grad_norm": 6.794834613800049, | |
| "learning_rate": 6.81992337164751e-05, | |
| "loss": 1.0533, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 13.551724137931034, | |
| "grad_norm": 6.377530097961426, | |
| "learning_rate": 6.808976464148879e-05, | |
| "loss": 0.8787, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 13.586206896551724, | |
| "grad_norm": 3.35634183883667, | |
| "learning_rate": 6.798029556650246e-05, | |
| "loss": 0.5949, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 13.620689655172415, | |
| "grad_norm": 2.987086057662964, | |
| "learning_rate": 6.787082649151615e-05, | |
| "loss": 0.7907, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 13.655172413793103, | |
| "grad_norm": 7.851985454559326, | |
| "learning_rate": 6.776135741652983e-05, | |
| "loss": 1.0953, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 13.689655172413794, | |
| "grad_norm": 5.5945892333984375, | |
| "learning_rate": 6.765188834154351e-05, | |
| "loss": 0.5258, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 13.724137931034482, | |
| "grad_norm": 6.0871052742004395, | |
| "learning_rate": 6.75424192665572e-05, | |
| "loss": 1.6107, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 13.758620689655173, | |
| "grad_norm": 4.6772260665893555, | |
| "learning_rate": 6.74329501915709e-05, | |
| "loss": 1.1545, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 13.793103448275861, | |
| "grad_norm": 2.2203543186187744, | |
| "learning_rate": 6.732348111658457e-05, | |
| "loss": 0.8467, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 13.827586206896552, | |
| "grad_norm": 5.075698375701904, | |
| "learning_rate": 6.721401204159825e-05, | |
| "loss": 1.047, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 13.862068965517242, | |
| "grad_norm": 9.27517032623291, | |
| "learning_rate": 6.710454296661194e-05, | |
| "loss": 0.9641, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 13.89655172413793, | |
| "grad_norm": 3.1357972621917725, | |
| "learning_rate": 6.699507389162561e-05, | |
| "loss": 0.6859, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 13.931034482758621, | |
| "grad_norm": 9.246620178222656, | |
| "learning_rate": 6.68856048166393e-05, | |
| "loss": 1.2339, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 13.96551724137931, | |
| "grad_norm": 4.054126739501953, | |
| "learning_rate": 6.677613574165298e-05, | |
| "loss": 0.888, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "grad_norm": 10.905935287475586, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 1.3729, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_accuracy": 0.6060606060606061, | |
| "eval_f1_macro": 0.5151332211371547, | |
| "eval_f1_micro": 0.6060606060606061, | |
| "eval_f1_weighted": 0.5860558713061843, | |
| "eval_loss": 1.2284520864486694, | |
| "eval_precision_macro": 0.5714167385740136, | |
| "eval_precision_micro": 0.6060606060606061, | |
| "eval_precision_weighted": 0.6313951407182923, | |
| "eval_recall_macro": 0.5224716553287982, | |
| "eval_recall_micro": 0.6060606060606061, | |
| "eval_recall_weighted": 0.6060606060606061, | |
| "eval_runtime": 1.7058, | |
| "eval_samples_per_second": 77.383, | |
| "eval_steps_per_second": 9.966, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 14.03448275862069, | |
| "grad_norm": 5.283136367797852, | |
| "learning_rate": 6.655719759168036e-05, | |
| "loss": 0.7426, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 14.068965517241379, | |
| "grad_norm": 3.3359534740448, | |
| "learning_rate": 6.644772851669404e-05, | |
| "loss": 0.5729, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 14.10344827586207, | |
| "grad_norm": 2.1584465503692627, | |
| "learning_rate": 6.633825944170772e-05, | |
| "loss": 0.6175, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 14.137931034482758, | |
| "grad_norm": 4.724003791809082, | |
| "learning_rate": 6.62287903667214e-05, | |
| "loss": 0.5157, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 14.172413793103448, | |
| "grad_norm": 6.147416591644287, | |
| "learning_rate": 6.611932129173509e-05, | |
| "loss": 0.7852, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 14.206896551724139, | |
| "grad_norm": 5.155529499053955, | |
| "learning_rate": 6.600985221674877e-05, | |
| "loss": 0.8213, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 14.241379310344827, | |
| "grad_norm": 2.3989200592041016, | |
| "learning_rate": 6.590038314176246e-05, | |
| "loss": 0.4571, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 14.275862068965518, | |
| "grad_norm": 4.3873090744018555, | |
| "learning_rate": 6.579091406677615e-05, | |
| "loss": 1.1529, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 14.310344827586206, | |
| "grad_norm": 5.521265029907227, | |
| "learning_rate": 6.568144499178982e-05, | |
| "loss": 0.841, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 14.344827586206897, | |
| "grad_norm": 3.919457197189331, | |
| "learning_rate": 6.55719759168035e-05, | |
| "loss": 0.6576, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 14.379310344827585, | |
| "grad_norm": 5.222126007080078, | |
| "learning_rate": 6.546250684181719e-05, | |
| "loss": 0.7117, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 14.413793103448276, | |
| "grad_norm": 5.21601676940918, | |
| "learning_rate": 6.535303776683088e-05, | |
| "loss": 0.6117, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 14.448275862068966, | |
| "grad_norm": 3.0854995250701904, | |
| "learning_rate": 6.524356869184456e-05, | |
| "loss": 0.4594, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 14.482758620689655, | |
| "grad_norm": 4.040173053741455, | |
| "learning_rate": 6.513409961685824e-05, | |
| "loss": 0.6008, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 14.517241379310345, | |
| "grad_norm": 8.294901847839355, | |
| "learning_rate": 6.502463054187192e-05, | |
| "loss": 1.344, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 14.551724137931034, | |
| "grad_norm": 3.9648568630218506, | |
| "learning_rate": 6.491516146688561e-05, | |
| "loss": 0.9206, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 14.586206896551724, | |
| "grad_norm": 5.604014873504639, | |
| "learning_rate": 6.480569239189928e-05, | |
| "loss": 0.9679, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 14.620689655172415, | |
| "grad_norm": 4.383110523223877, | |
| "learning_rate": 6.469622331691297e-05, | |
| "loss": 0.8063, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 14.655172413793103, | |
| "grad_norm": 3.8494396209716797, | |
| "learning_rate": 6.458675424192667e-05, | |
| "loss": 1.0672, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 14.689655172413794, | |
| "grad_norm": 2.571416139602661, | |
| "learning_rate": 6.447728516694034e-05, | |
| "loss": 0.7218, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 14.724137931034482, | |
| "grad_norm": 13.170735359191895, | |
| "learning_rate": 6.436781609195403e-05, | |
| "loss": 0.7612, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 14.758620689655173, | |
| "grad_norm": 4.889159202575684, | |
| "learning_rate": 6.425834701696771e-05, | |
| "loss": 0.7591, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 14.793103448275861, | |
| "grad_norm": 4.082027435302734, | |
| "learning_rate": 6.414887794198139e-05, | |
| "loss": 0.9275, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 14.827586206896552, | |
| "grad_norm": 4.967465400695801, | |
| "learning_rate": 6.403940886699507e-05, | |
| "loss": 0.4512, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 14.862068965517242, | |
| "grad_norm": 8.991899490356445, | |
| "learning_rate": 6.392993979200876e-05, | |
| "loss": 0.8968, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 14.89655172413793, | |
| "grad_norm": 3.955719232559204, | |
| "learning_rate": 6.382047071702244e-05, | |
| "loss": 0.4046, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 14.931034482758621, | |
| "grad_norm": 8.865561485290527, | |
| "learning_rate": 6.371100164203613e-05, | |
| "loss": 0.8692, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 14.96551724137931, | |
| "grad_norm": 4.807842254638672, | |
| "learning_rate": 6.360153256704982e-05, | |
| "loss": 1.1433, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 3.4882490634918213, | |
| "learning_rate": 6.349206349206349e-05, | |
| "loss": 1.3638, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_accuracy": 0.6212121212121212, | |
| "eval_f1_macro": 0.5388922283385812, | |
| "eval_f1_micro": 0.6212121212121212, | |
| "eval_f1_weighted": 0.6054871995244131, | |
| "eval_loss": 1.1741819381713867, | |
| "eval_precision_macro": 0.561660793349598, | |
| "eval_precision_micro": 0.6212121212121212, | |
| "eval_precision_weighted": 0.6333975796499517, | |
| "eval_recall_macro": 0.5513378684807256, | |
| "eval_recall_micro": 0.6212121212121212, | |
| "eval_recall_weighted": 0.6212121212121212, | |
| "eval_runtime": 1.6952, | |
| "eval_samples_per_second": 77.867, | |
| "eval_steps_per_second": 10.028, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 15.03448275862069, | |
| "grad_norm": 6.959043502807617, | |
| "learning_rate": 6.338259441707718e-05, | |
| "loss": 0.6661, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 15.068965517241379, | |
| "grad_norm": 4.410765647888184, | |
| "learning_rate": 6.327312534209086e-05, | |
| "loss": 0.6515, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 15.10344827586207, | |
| "grad_norm": 5.796648025512695, | |
| "learning_rate": 6.316365626710455e-05, | |
| "loss": 0.5223, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 15.137931034482758, | |
| "grad_norm": 7.7657904624938965, | |
| "learning_rate": 6.305418719211823e-05, | |
| "loss": 0.6789, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 15.172413793103448, | |
| "grad_norm": 2.134707450866699, | |
| "learning_rate": 6.294471811713192e-05, | |
| "loss": 0.3882, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 15.206896551724139, | |
| "grad_norm": 2.849294900894165, | |
| "learning_rate": 6.283524904214559e-05, | |
| "loss": 0.8977, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 15.241379310344827, | |
| "grad_norm": 3.2845396995544434, | |
| "learning_rate": 6.272577996715928e-05, | |
| "loss": 0.5675, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 15.275862068965518, | |
| "grad_norm": 7.925327777862549, | |
| "learning_rate": 6.261631089217297e-05, | |
| "loss": 1.0438, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 15.310344827586206, | |
| "grad_norm": 3.015355348587036, | |
| "learning_rate": 6.250684181718665e-05, | |
| "loss": 0.6477, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 15.344827586206897, | |
| "grad_norm": 4.829744815826416, | |
| "learning_rate": 6.239737274220034e-05, | |
| "loss": 0.8828, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 15.379310344827585, | |
| "grad_norm": 1.1450001001358032, | |
| "learning_rate": 6.228790366721402e-05, | |
| "loss": 0.59, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 15.413793103448276, | |
| "grad_norm": 3.852428436279297, | |
| "learning_rate": 6.21784345922277e-05, | |
| "loss": 0.6749, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 15.448275862068966, | |
| "grad_norm": 8.538708686828613, | |
| "learning_rate": 6.206896551724138e-05, | |
| "loss": 0.7512, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 15.482758620689655, | |
| "grad_norm": 1.829010248184204, | |
| "learning_rate": 6.195949644225506e-05, | |
| "loss": 0.5363, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 15.517241379310345, | |
| "grad_norm": 4.311962127685547, | |
| "learning_rate": 6.185002736726874e-05, | |
| "loss": 0.507, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 15.551724137931034, | |
| "grad_norm": 6.414646625518799, | |
| "learning_rate": 6.174055829228244e-05, | |
| "loss": 0.4162, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 15.586206896551724, | |
| "grad_norm": 4.423099517822266, | |
| "learning_rate": 6.163108921729611e-05, | |
| "loss": 0.8151, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 15.620689655172415, | |
| "grad_norm": 1.6412297487258911, | |
| "learning_rate": 6.15216201423098e-05, | |
| "loss": 0.4281, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 15.655172413793103, | |
| "grad_norm": 8.947938919067383, | |
| "learning_rate": 6.141215106732349e-05, | |
| "loss": 0.8403, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 15.689655172413794, | |
| "grad_norm": 8.341460227966309, | |
| "learning_rate": 6.130268199233716e-05, | |
| "loss": 0.7946, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 15.724137931034482, | |
| "grad_norm": 7.36137580871582, | |
| "learning_rate": 6.119321291735085e-05, | |
| "loss": 0.3792, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 15.758620689655173, | |
| "grad_norm": 4.560554504394531, | |
| "learning_rate": 6.108374384236453e-05, | |
| "loss": 0.7343, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 15.793103448275861, | |
| "grad_norm": 10.598052978515625, | |
| "learning_rate": 6.097427476737821e-05, | |
| "loss": 0.6625, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 15.827586206896552, | |
| "grad_norm": 3.1005873680114746, | |
| "learning_rate": 6.0864805692391905e-05, | |
| "loss": 0.4325, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 15.862068965517242, | |
| "grad_norm": 4.4200663566589355, | |
| "learning_rate": 6.075533661740559e-05, | |
| "loss": 0.9942, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 15.89655172413793, | |
| "grad_norm": 5.5699381828308105, | |
| "learning_rate": 6.0645867542419264e-05, | |
| "loss": 0.3725, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 15.931034482758621, | |
| "grad_norm": 4.177437782287598, | |
| "learning_rate": 6.053639846743295e-05, | |
| "loss": 0.5522, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 15.96551724137931, | |
| "grad_norm": 2.993516206741333, | |
| "learning_rate": 6.042692939244664e-05, | |
| "loss": 0.4704, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 4.057467937469482, | |
| "learning_rate": 6.0317460317460316e-05, | |
| "loss": 0.9063, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_accuracy": 0.5984848484848485, | |
| "eval_f1_macro": 0.5079468662671953, | |
| "eval_f1_micro": 0.5984848484848485, | |
| "eval_f1_weighted": 0.5770103503469938, | |
| "eval_loss": 1.2325303554534912, | |
| "eval_precision_macro": 0.50772811148751, | |
| "eval_precision_micro": 0.5984848484848485, | |
| "eval_precision_weighted": 0.5714871385924017, | |
| "eval_recall_macro": 0.52151171579743, | |
| "eval_recall_micro": 0.5984848484848485, | |
| "eval_recall_weighted": 0.5984848484848485, | |
| "eval_runtime": 1.6873, | |
| "eval_samples_per_second": 78.231, | |
| "eval_steps_per_second": 10.075, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 16.03448275862069, | |
| "grad_norm": 1.7874680757522583, | |
| "learning_rate": 6.0207991242474e-05, | |
| "loss": 0.2892, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 16.06896551724138, | |
| "grad_norm": 3.2080960273742676, | |
| "learning_rate": 6.0098522167487695e-05, | |
| "loss": 0.7073, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 16.103448275862068, | |
| "grad_norm": 1.299979567527771, | |
| "learning_rate": 5.998905309250137e-05, | |
| "loss": 0.2762, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 16.137931034482758, | |
| "grad_norm": 2.3571527004241943, | |
| "learning_rate": 5.9879584017515054e-05, | |
| "loss": 0.5417, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 16.17241379310345, | |
| "grad_norm": 6.40372371673584, | |
| "learning_rate": 5.977011494252874e-05, | |
| "loss": 0.8834, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 16.20689655172414, | |
| "grad_norm": 3.0044682025909424, | |
| "learning_rate": 5.966064586754242e-05, | |
| "loss": 0.3184, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 16.24137931034483, | |
| "grad_norm": 10.370132446289062, | |
| "learning_rate": 5.9551176792556106e-05, | |
| "loss": 0.8784, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 16.275862068965516, | |
| "grad_norm": 4.730606555938721, | |
| "learning_rate": 5.944170771756979e-05, | |
| "loss": 0.8115, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 16.310344827586206, | |
| "grad_norm": 8.03954029083252, | |
| "learning_rate": 5.933223864258347e-05, | |
| "loss": 0.4177, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 16.344827586206897, | |
| "grad_norm": 4.975473403930664, | |
| "learning_rate": 5.922276956759716e-05, | |
| "loss": 0.4525, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 16.379310344827587, | |
| "grad_norm": 4.188839912414551, | |
| "learning_rate": 5.9113300492610844e-05, | |
| "loss": 0.8818, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 16.413793103448278, | |
| "grad_norm": 7.270519733428955, | |
| "learning_rate": 5.900383141762452e-05, | |
| "loss": 0.6999, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 16.448275862068964, | |
| "grad_norm": 3.524183988571167, | |
| "learning_rate": 5.889436234263821e-05, | |
| "loss": 0.9776, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 16.482758620689655, | |
| "grad_norm": 3.496511936187744, | |
| "learning_rate": 5.878489326765189e-05, | |
| "loss": 0.6665, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 16.517241379310345, | |
| "grad_norm": 8.359578132629395, | |
| "learning_rate": 5.8675424192665575e-05, | |
| "loss": 0.5769, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 16.551724137931036, | |
| "grad_norm": 1.640060305595398, | |
| "learning_rate": 5.856595511767926e-05, | |
| "loss": 0.4558, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 16.586206896551722, | |
| "grad_norm": 3.3751068115234375, | |
| "learning_rate": 5.8456486042692934e-05, | |
| "loss": 0.7395, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 16.620689655172413, | |
| "grad_norm": 6.145583629608154, | |
| "learning_rate": 5.834701696770663e-05, | |
| "loss": 0.5962, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 16.655172413793103, | |
| "grad_norm": 4.918842315673828, | |
| "learning_rate": 5.823754789272031e-05, | |
| "loss": 0.9492, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 16.689655172413794, | |
| "grad_norm": 8.429754257202148, | |
| "learning_rate": 5.8128078817733986e-05, | |
| "loss": 0.7788, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 16.724137931034484, | |
| "grad_norm": 1.9383467435836792, | |
| "learning_rate": 5.801860974274768e-05, | |
| "loss": 0.561, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 16.75862068965517, | |
| "grad_norm": 8.061017990112305, | |
| "learning_rate": 5.7909140667761365e-05, | |
| "loss": 0.5616, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 16.79310344827586, | |
| "grad_norm": 3.3627376556396484, | |
| "learning_rate": 5.779967159277504e-05, | |
| "loss": 0.4317, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 16.82758620689655, | |
| "grad_norm": 4.939265727996826, | |
| "learning_rate": 5.7690202517788724e-05, | |
| "loss": 0.7131, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 16.862068965517242, | |
| "grad_norm": 8.868173599243164, | |
| "learning_rate": 5.758073344280242e-05, | |
| "loss": 0.848, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 16.896551724137932, | |
| "grad_norm": 6.587118625640869, | |
| "learning_rate": 5.747126436781609e-05, | |
| "loss": 0.4767, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 16.93103448275862, | |
| "grad_norm": 5.797451972961426, | |
| "learning_rate": 5.7361795292829776e-05, | |
| "loss": 0.625, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 16.96551724137931, | |
| "grad_norm": 3.9016692638397217, | |
| "learning_rate": 5.725232621784347e-05, | |
| "loss": 0.6723, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "grad_norm": 2.006951332092285, | |
| "learning_rate": 5.714285714285714e-05, | |
| "loss": 0.4584, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_accuracy": 0.6363636363636364, | |
| "eval_f1_macro": 0.5514969546015317, | |
| "eval_f1_micro": 0.6363636363636364, | |
| "eval_f1_weighted": 0.6209731762304284, | |
| "eval_loss": 1.1496680974960327, | |
| "eval_precision_macro": 0.5676373113347903, | |
| "eval_precision_micro": 0.6363636363636364, | |
| "eval_precision_weighted": 0.6285885138826316, | |
| "eval_recall_macro": 0.5575207860922147, | |
| "eval_recall_micro": 0.6363636363636364, | |
| "eval_recall_weighted": 0.6363636363636364, | |
| "eval_runtime": 1.699, | |
| "eval_samples_per_second": 77.694, | |
| "eval_steps_per_second": 10.006, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 17.03448275862069, | |
| "grad_norm": 6.7209343910217285, | |
| "learning_rate": 5.703338806787083e-05, | |
| "loss": 0.5839, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 17.06896551724138, | |
| "grad_norm": 4.373691558837891, | |
| "learning_rate": 5.6923918992884514e-05, | |
| "loss": 0.455, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 17.103448275862068, | |
| "grad_norm": 7.231139183044434, | |
| "learning_rate": 5.6814449917898194e-05, | |
| "loss": 1.0114, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 17.137931034482758, | |
| "grad_norm": 7.747330665588379, | |
| "learning_rate": 5.670498084291188e-05, | |
| "loss": 0.9266, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 17.17241379310345, | |
| "grad_norm": 6.351534843444824, | |
| "learning_rate": 5.6595511767925566e-05, | |
| "loss": 0.4284, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 17.20689655172414, | |
| "grad_norm": 10.491331100463867, | |
| "learning_rate": 5.6486042692939246e-05, | |
| "loss": 0.4253, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 17.24137931034483, | |
| "grad_norm": 4.956140518188477, | |
| "learning_rate": 5.637657361795293e-05, | |
| "loss": 0.4558, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 17.275862068965516, | |
| "grad_norm": 3.0368525981903076, | |
| "learning_rate": 5.626710454296662e-05, | |
| "loss": 1.0867, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 17.310344827586206, | |
| "grad_norm": 1.7548032999038696, | |
| "learning_rate": 5.61576354679803e-05, | |
| "loss": 0.2738, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 17.344827586206897, | |
| "grad_norm": 3.532531261444092, | |
| "learning_rate": 5.6048166392993984e-05, | |
| "loss": 0.344, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 17.379310344827587, | |
| "grad_norm": 8.146524429321289, | |
| "learning_rate": 5.593869731800766e-05, | |
| "loss": 0.67, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 17.413793103448278, | |
| "grad_norm": 11.48868465423584, | |
| "learning_rate": 5.582922824302135e-05, | |
| "loss": 0.5444, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 17.448275862068964, | |
| "grad_norm": 4.138803958892822, | |
| "learning_rate": 5.5719759168035036e-05, | |
| "loss": 0.4858, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 17.482758620689655, | |
| "grad_norm": 2.8997292518615723, | |
| "learning_rate": 5.561029009304871e-05, | |
| "loss": 0.2196, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 17.517241379310345, | |
| "grad_norm": 7.666640281677246, | |
| "learning_rate": 5.55008210180624e-05, | |
| "loss": 0.7209, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 17.551724137931036, | |
| "grad_norm": 1.5840152502059937, | |
| "learning_rate": 5.539135194307609e-05, | |
| "loss": 0.4648, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 17.586206896551722, | |
| "grad_norm": 3.6097986698150635, | |
| "learning_rate": 5.528188286808976e-05, | |
| "loss": 0.5552, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 17.620689655172413, | |
| "grad_norm": 9.492504119873047, | |
| "learning_rate": 5.517241379310345e-05, | |
| "loss": 0.9165, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 17.655172413793103, | |
| "grad_norm": 8.897369384765625, | |
| "learning_rate": 5.506294471811714e-05, | |
| "loss": 0.5663, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 17.689655172413794, | |
| "grad_norm": 3.185572862625122, | |
| "learning_rate": 5.495347564313081e-05, | |
| "loss": 0.5031, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 17.724137931034484, | |
| "grad_norm": 2.46030855178833, | |
| "learning_rate": 5.48440065681445e-05, | |
| "loss": 0.5205, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 17.75862068965517, | |
| "grad_norm": 1.1284418106079102, | |
| "learning_rate": 5.473453749315819e-05, | |
| "loss": 0.5335, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 17.79310344827586, | |
| "grad_norm": 7.253750801086426, | |
| "learning_rate": 5.4625068418171864e-05, | |
| "loss": 0.655, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 17.82758620689655, | |
| "grad_norm": 3.3334109783172607, | |
| "learning_rate": 5.451559934318555e-05, | |
| "loss": 0.3426, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 17.862068965517242, | |
| "grad_norm": 4.30219841003418, | |
| "learning_rate": 5.440613026819924e-05, | |
| "loss": 0.5999, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 17.896551724137932, | |
| "grad_norm": 6.265329360961914, | |
| "learning_rate": 5.4296661193212916e-05, | |
| "loss": 0.4511, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 17.93103448275862, | |
| "grad_norm": 1.11660635471344, | |
| "learning_rate": 5.41871921182266e-05, | |
| "loss": 0.4022, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 17.96551724137931, | |
| "grad_norm": 4.282124996185303, | |
| "learning_rate": 5.407772304324029e-05, | |
| "loss": 0.3705, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "grad_norm": 4.951636791229248, | |
| "learning_rate": 5.396825396825397e-05, | |
| "loss": 0.86, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_accuracy": 0.5909090909090909, | |
| "eval_f1_macro": 0.49254721886300834, | |
| "eval_f1_micro": 0.5909090909090909, | |
| "eval_f1_weighted": 0.5718522484072723, | |
| "eval_loss": 1.2673066854476929, | |
| "eval_precision_macro": 0.49684601113172544, | |
| "eval_precision_micro": 0.5909090909090909, | |
| "eval_precision_weighted": 0.568108028335301, | |
| "eval_recall_macro": 0.5030990173847316, | |
| "eval_recall_micro": 0.5909090909090909, | |
| "eval_recall_weighted": 0.5909090909090909, | |
| "eval_runtime": 1.6855, | |
| "eval_samples_per_second": 78.316, | |
| "eval_steps_per_second": 10.086, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 18.03448275862069, | |
| "grad_norm": 5.494875907897949, | |
| "learning_rate": 5.3858784893267654e-05, | |
| "loss": 0.6363, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 18.06896551724138, | |
| "grad_norm": 5.767899036407471, | |
| "learning_rate": 5.374931581828134e-05, | |
| "loss": 0.2099, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 18.103448275862068, | |
| "grad_norm": 2.901460886001587, | |
| "learning_rate": 5.363984674329502e-05, | |
| "loss": 0.2682, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 18.137931034482758, | |
| "grad_norm": 7.761710166931152, | |
| "learning_rate": 5.3530377668308706e-05, | |
| "loss": 0.3704, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 18.17241379310345, | |
| "grad_norm": 4.4380879402160645, | |
| "learning_rate": 5.342090859332239e-05, | |
| "loss": 0.273, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 18.20689655172414, | |
| "grad_norm": 7.681371688842773, | |
| "learning_rate": 5.331143951833607e-05, | |
| "loss": 0.5835, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 18.24137931034483, | |
| "grad_norm": 10.872917175292969, | |
| "learning_rate": 5.320197044334976e-05, | |
| "loss": 0.529, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 18.275862068965516, | |
| "grad_norm": 6.020653247833252, | |
| "learning_rate": 5.3092501368363444e-05, | |
| "loss": 0.3757, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 18.310344827586206, | |
| "grad_norm": 7.052387714385986, | |
| "learning_rate": 5.2983032293377124e-05, | |
| "loss": 0.3641, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 18.344827586206897, | |
| "grad_norm": 6.335596084594727, | |
| "learning_rate": 5.287356321839081e-05, | |
| "loss": 0.3232, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 18.379310344827587, | |
| "grad_norm": 6.076163291931152, | |
| "learning_rate": 5.276409414340448e-05, | |
| "loss": 0.6439, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 18.413793103448278, | |
| "grad_norm": 0.6590009331703186, | |
| "learning_rate": 5.2654625068418176e-05, | |
| "loss": 0.1923, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 18.448275862068964, | |
| "grad_norm": 0.6382133364677429, | |
| "learning_rate": 5.254515599343186e-05, | |
| "loss": 0.567, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 18.482758620689655, | |
| "grad_norm": 10.746706008911133, | |
| "learning_rate": 5.2435686918445535e-05, | |
| "loss": 0.6355, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 18.517241379310345, | |
| "grad_norm": 6.357698917388916, | |
| "learning_rate": 5.232621784345923e-05, | |
| "loss": 0.4048, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 18.551724137931036, | |
| "grad_norm": 4.186829566955566, | |
| "learning_rate": 5.2216748768472914e-05, | |
| "loss": 0.5531, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 18.586206896551722, | |
| "grad_norm": 7.867221832275391, | |
| "learning_rate": 5.2107279693486586e-05, | |
| "loss": 0.3116, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 18.620689655172413, | |
| "grad_norm": 3.410994052886963, | |
| "learning_rate": 5.199781061850027e-05, | |
| "loss": 0.4549, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 18.655172413793103, | |
| "grad_norm": 4.2880706787109375, | |
| "learning_rate": 5.1888341543513966e-05, | |
| "loss": 0.5768, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 18.689655172413794, | |
| "grad_norm": 5.609067440032959, | |
| "learning_rate": 5.177887246852764e-05, | |
| "loss": 0.2996, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 18.724137931034484, | |
| "grad_norm": 7.243040561676025, | |
| "learning_rate": 5.1669403393541325e-05, | |
| "loss": 0.6335, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 18.75862068965517, | |
| "grad_norm": 9.456660270690918, | |
| "learning_rate": 5.155993431855502e-05, | |
| "loss": 0.771, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 18.79310344827586, | |
| "grad_norm": 3.2803101539611816, | |
| "learning_rate": 5.145046524356869e-05, | |
| "loss": 0.4275, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 18.82758620689655, | |
| "grad_norm": 5.173320293426514, | |
| "learning_rate": 5.1340996168582377e-05, | |
| "loss": 0.5103, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 18.862068965517242, | |
| "grad_norm": 9.215474128723145, | |
| "learning_rate": 5.123152709359606e-05, | |
| "loss": 0.4572, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 18.896551724137932, | |
| "grad_norm": 3.3474466800689697, | |
| "learning_rate": 5.112205801860974e-05, | |
| "loss": 0.1896, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 18.93103448275862, | |
| "grad_norm": 4.091315746307373, | |
| "learning_rate": 5.101258894362343e-05, | |
| "loss": 0.6855, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 18.96551724137931, | |
| "grad_norm": 5.4540696144104, | |
| "learning_rate": 5.0903119868637115e-05, | |
| "loss": 0.2874, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "grad_norm": 0.8723583817481995, | |
| "learning_rate": 5.0793650793650794e-05, | |
| "loss": 0.2113, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_accuracy": 0.6212121212121212, | |
| "eval_f1_macro": 0.5179873730419885, | |
| "eval_f1_micro": 0.6212121212121212, | |
| "eval_f1_weighted": 0.5986162885696074, | |
| "eval_loss": 1.2132326364517212, | |
| "eval_precision_macro": 0.5386071190769648, | |
| "eval_precision_micro": 0.6212121212121212, | |
| "eval_precision_weighted": 0.6049018570893172, | |
| "eval_recall_macro": 0.5257445200302343, | |
| "eval_recall_micro": 0.6212121212121212, | |
| "eval_recall_weighted": 0.6212121212121212, | |
| "eval_runtime": 1.7051, | |
| "eval_samples_per_second": 77.414, | |
| "eval_steps_per_second": 9.97, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 19.03448275862069, | |
| "grad_norm": 3.2095038890838623, | |
| "learning_rate": 5.068418171866448e-05, | |
| "loss": 0.3305, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 19.06896551724138, | |
| "grad_norm": 7.022287368774414, | |
| "learning_rate": 5.057471264367817e-05, | |
| "loss": 0.6748, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 19.103448275862068, | |
| "grad_norm": 8.976895332336426, | |
| "learning_rate": 5.0465243568691846e-05, | |
| "loss": 0.4676, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 19.137931034482758, | |
| "grad_norm": 5.533504962921143, | |
| "learning_rate": 5.035577449370553e-05, | |
| "loss": 0.2232, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 19.17241379310345, | |
| "grad_norm": 4.6383466720581055, | |
| "learning_rate": 5.024630541871922e-05, | |
| "loss": 0.4356, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 19.20689655172414, | |
| "grad_norm": 0.5434678792953491, | |
| "learning_rate": 5.01368363437329e-05, | |
| "loss": 0.1359, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 19.24137931034483, | |
| "grad_norm": 4.4346795082092285, | |
| "learning_rate": 5.0027367268746584e-05, | |
| "loss": 0.5523, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 19.275862068965516, | |
| "grad_norm": 2.3363232612609863, | |
| "learning_rate": 4.9917898193760264e-05, | |
| "loss": 0.1374, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 19.310344827586206, | |
| "grad_norm": 3.592006206512451, | |
| "learning_rate": 4.980842911877395e-05, | |
| "loss": 0.4274, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 19.344827586206897, | |
| "grad_norm": 1.3552830219268799, | |
| "learning_rate": 4.9698960043787636e-05, | |
| "loss": 0.2087, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 19.379310344827587, | |
| "grad_norm": 5.443765640258789, | |
| "learning_rate": 4.9589490968801316e-05, | |
| "loss": 0.3383, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 19.413793103448278, | |
| "grad_norm": 8.5775785446167, | |
| "learning_rate": 4.9480021893814995e-05, | |
| "loss": 0.3641, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 19.448275862068964, | |
| "grad_norm": 8.10037899017334, | |
| "learning_rate": 4.937055281882869e-05, | |
| "loss": 0.4428, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 19.482758620689655, | |
| "grad_norm": 1.8419495820999146, | |
| "learning_rate": 4.926108374384237e-05, | |
| "loss": 0.4689, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 19.517241379310345, | |
| "grad_norm": 8.350903511047363, | |
| "learning_rate": 4.915161466885605e-05, | |
| "loss": 0.1867, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 19.551724137931036, | |
| "grad_norm": 0.38135936856269836, | |
| "learning_rate": 4.904214559386973e-05, | |
| "loss": 0.2662, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 19.586206896551722, | |
| "grad_norm": 7.118171691894531, | |
| "learning_rate": 4.893267651888342e-05, | |
| "loss": 0.4822, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 19.620689655172413, | |
| "grad_norm": 6.289510250091553, | |
| "learning_rate": 4.88232074438971e-05, | |
| "loss": 0.3416, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 19.655172413793103, | |
| "grad_norm": 7.544029235839844, | |
| "learning_rate": 4.8713738368910785e-05, | |
| "loss": 0.2186, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 19.689655172413794, | |
| "grad_norm": 5.812250137329102, | |
| "learning_rate": 4.860426929392447e-05, | |
| "loss": 0.9843, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 19.724137931034484, | |
| "grad_norm": 5.986517429351807, | |
| "learning_rate": 4.849480021893815e-05, | |
| "loss": 0.4633, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 19.75862068965517, | |
| "grad_norm": 11.750970840454102, | |
| "learning_rate": 4.838533114395184e-05, | |
| "loss": 0.2648, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 19.79310344827586, | |
| "grad_norm": 1.8454886674880981, | |
| "learning_rate": 4.827586206896552e-05, | |
| "loss": 0.3346, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 19.82758620689655, | |
| "grad_norm": 12.597700119018555, | |
| "learning_rate": 4.81663929939792e-05, | |
| "loss": 0.54, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 19.862068965517242, | |
| "grad_norm": 8.612327575683594, | |
| "learning_rate": 4.805692391899288e-05, | |
| "loss": 0.3575, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 19.896551724137932, | |
| "grad_norm": 3.5571696758270264, | |
| "learning_rate": 4.7947454844006575e-05, | |
| "loss": 0.5335, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 19.93103448275862, | |
| "grad_norm": 5.682160377502441, | |
| "learning_rate": 4.7837985769020255e-05, | |
| "loss": 0.2374, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 19.96551724137931, | |
| "grad_norm": 6.373334884643555, | |
| "learning_rate": 4.7728516694033934e-05, | |
| "loss": 0.6124, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 0.5094832181930542, | |
| "learning_rate": 4.761904761904762e-05, | |
| "loss": 0.1168, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 0.6136363636363636, | |
| "eval_f1_macro": 0.5543005733250872, | |
| "eval_f1_micro": 0.6136363636363636, | |
| "eval_f1_weighted": 0.6069833060571026, | |
| "eval_loss": 1.2442402839660645, | |
| "eval_precision_macro": 0.5742089820117771, | |
| "eval_precision_micro": 0.6136363636363636, | |
| "eval_precision_weighted": 0.616364865879877, | |
| "eval_recall_macro": 0.5517233560090703, | |
| "eval_recall_micro": 0.6136363636363636, | |
| "eval_recall_weighted": 0.6136363636363636, | |
| "eval_runtime": 1.7788, | |
| "eval_samples_per_second": 74.206, | |
| "eval_steps_per_second": 9.557, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 20.03448275862069, | |
| "grad_norm": 4.78882360458374, | |
| "learning_rate": 4.7509578544061307e-05, | |
| "loss": 0.142, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 20.06896551724138, | |
| "grad_norm": 9.2319917678833, | |
| "learning_rate": 4.7400109469074986e-05, | |
| "loss": 0.2529, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 20.103448275862068, | |
| "grad_norm": 9.061631202697754, | |
| "learning_rate": 4.729064039408867e-05, | |
| "loss": 0.6863, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 20.137931034482758, | |
| "grad_norm": 2.0638833045959473, | |
| "learning_rate": 4.718117131910236e-05, | |
| "loss": 0.4872, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 20.17241379310345, | |
| "grad_norm": 1.029740571975708, | |
| "learning_rate": 4.707170224411604e-05, | |
| "loss": 0.2375, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 20.20689655172414, | |
| "grad_norm": 4.14223575592041, | |
| "learning_rate": 4.6962233169129724e-05, | |
| "loss": 0.6973, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 20.24137931034483, | |
| "grad_norm": 5.068900108337402, | |
| "learning_rate": 4.685276409414341e-05, | |
| "loss": 0.7413, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 20.275862068965516, | |
| "grad_norm": 2.607081413269043, | |
| "learning_rate": 4.674329501915709e-05, | |
| "loss": 0.454, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 20.310344827586206, | |
| "grad_norm": 2.115675926208496, | |
| "learning_rate": 4.663382594417077e-05, | |
| "loss": 0.3017, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 20.344827586206897, | |
| "grad_norm": 1.3065847158432007, | |
| "learning_rate": 4.652435686918446e-05, | |
| "loss": 0.4131, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 20.379310344827587, | |
| "grad_norm": 10.156610488891602, | |
| "learning_rate": 4.641488779419814e-05, | |
| "loss": 0.6956, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 20.413793103448278, | |
| "grad_norm": 6.317163944244385, | |
| "learning_rate": 4.630541871921182e-05, | |
| "loss": 0.1953, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 20.448275862068964, | |
| "grad_norm": 4.764923095703125, | |
| "learning_rate": 4.619594964422551e-05, | |
| "loss": 0.1456, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 20.482758620689655, | |
| "grad_norm": 3.1632251739501953, | |
| "learning_rate": 4.6086480569239194e-05, | |
| "loss": 0.2287, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 20.517241379310345, | |
| "grad_norm": 12.108798027038574, | |
| "learning_rate": 4.597701149425287e-05, | |
| "loss": 0.2027, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 20.551724137931036, | |
| "grad_norm": 11.341327667236328, | |
| "learning_rate": 4.586754241926656e-05, | |
| "loss": 0.4052, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 20.586206896551722, | |
| "grad_norm": 5.067739009857178, | |
| "learning_rate": 4.5758073344280246e-05, | |
| "loss": 0.1575, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 20.620689655172413, | |
| "grad_norm": 7.721131324768066, | |
| "learning_rate": 4.5648604269293925e-05, | |
| "loss": 0.4256, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 20.655172413793103, | |
| "grad_norm": 10.028607368469238, | |
| "learning_rate": 4.553913519430761e-05, | |
| "loss": 0.5147, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 20.689655172413794, | |
| "grad_norm": 1.6635549068450928, | |
| "learning_rate": 4.54296661193213e-05, | |
| "loss": 0.1427, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 20.724137931034484, | |
| "grad_norm": 6.918335914611816, | |
| "learning_rate": 4.532019704433498e-05, | |
| "loss": 0.2659, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 20.75862068965517, | |
| "grad_norm": 2.401365041732788, | |
| "learning_rate": 4.5210727969348656e-05, | |
| "loss": 0.3284, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 20.79310344827586, | |
| "grad_norm": 3.1419484615325928, | |
| "learning_rate": 4.510125889436235e-05, | |
| "loss": 0.2873, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 20.82758620689655, | |
| "grad_norm": 5.024045944213867, | |
| "learning_rate": 4.499178981937603e-05, | |
| "loss": 0.3941, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 20.862068965517242, | |
| "grad_norm": 0.6695942878723145, | |
| "learning_rate": 4.488232074438971e-05, | |
| "loss": 0.103, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 20.896551724137932, | |
| "grad_norm": 2.601149320602417, | |
| "learning_rate": 4.47728516694034e-05, | |
| "loss": 0.2325, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 20.93103448275862, | |
| "grad_norm": 10.553409576416016, | |
| "learning_rate": 4.466338259441708e-05, | |
| "loss": 0.3237, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 20.96551724137931, | |
| "grad_norm": 4.566164016723633, | |
| "learning_rate": 4.455391351943076e-05, | |
| "loss": 0.2796, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "grad_norm": 0.7345550060272217, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "loss": 0.3149, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5445690340153033, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6145913352521732, | |
| "eval_loss": 1.2900217771530151, | |
| "eval_precision_macro": 0.5463030249320572, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6119575429051235, | |
| "eval_recall_macro": 0.5533938019652305, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 1.6672, | |
| "eval_samples_per_second": 79.174, | |
| "eval_steps_per_second": 10.197, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 21.03448275862069, | |
| "grad_norm": 6.113945007324219, | |
| "learning_rate": 4.433497536945813e-05, | |
| "loss": 0.1953, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 21.06896551724138, | |
| "grad_norm": 6.228808879852295, | |
| "learning_rate": 4.422550629447181e-05, | |
| "loss": 0.1327, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 21.103448275862068, | |
| "grad_norm": 2.212075710296631, | |
| "learning_rate": 4.41160372194855e-05, | |
| "loss": 0.0969, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 21.137931034482758, | |
| "grad_norm": 0.48354583978652954, | |
| "learning_rate": 4.4006568144499185e-05, | |
| "loss": 0.3936, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 21.17241379310345, | |
| "grad_norm": 20.688846588134766, | |
| "learning_rate": 4.3897099069512864e-05, | |
| "loss": 0.5529, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 21.20689655172414, | |
| "grad_norm": 0.4432692527770996, | |
| "learning_rate": 4.3787629994526544e-05, | |
| "loss": 0.2458, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 21.24137931034483, | |
| "grad_norm": 0.3409382402896881, | |
| "learning_rate": 4.367816091954024e-05, | |
| "loss": 0.0771, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 21.275862068965516, | |
| "grad_norm": 3.208294153213501, | |
| "learning_rate": 4.3568691844553916e-05, | |
| "loss": 0.1512, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 21.310344827586206, | |
| "grad_norm": 5.407877445220947, | |
| "learning_rate": 4.3459222769567596e-05, | |
| "loss": 0.1977, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 21.344827586206897, | |
| "grad_norm": 6.478481292724609, | |
| "learning_rate": 4.334975369458129e-05, | |
| "loss": 0.3786, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 21.379310344827587, | |
| "grad_norm": 12.132913589477539, | |
| "learning_rate": 4.324028461959497e-05, | |
| "loss": 0.2917, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 21.413793103448278, | |
| "grad_norm": 3.9148740768432617, | |
| "learning_rate": 4.313081554460865e-05, | |
| "loss": 0.2521, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 21.448275862068964, | |
| "grad_norm": 9.572196006774902, | |
| "learning_rate": 4.3021346469622334e-05, | |
| "loss": 0.5989, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 21.482758620689655, | |
| "grad_norm": 14.302295684814453, | |
| "learning_rate": 4.291187739463602e-05, | |
| "loss": 0.3296, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 21.517241379310345, | |
| "grad_norm": 3.571140766143799, | |
| "learning_rate": 4.28024083196497e-05, | |
| "loss": 0.2097, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 21.551724137931036, | |
| "grad_norm": 20.992183685302734, | |
| "learning_rate": 4.2692939244663386e-05, | |
| "loss": 0.4902, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 21.586206896551722, | |
| "grad_norm": 0.8532655239105225, | |
| "learning_rate": 4.258347016967707e-05, | |
| "loss": 0.3269, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 21.620689655172413, | |
| "grad_norm": 6.140798091888428, | |
| "learning_rate": 4.247400109469075e-05, | |
| "loss": 0.3355, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 21.655172413793103, | |
| "grad_norm": 3.988628625869751, | |
| "learning_rate": 4.236453201970443e-05, | |
| "loss": 0.3815, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 21.689655172413794, | |
| "grad_norm": 3.6218347549438477, | |
| "learning_rate": 4.2255062944718124e-05, | |
| "loss": 0.3056, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 21.724137931034484, | |
| "grad_norm": 2.697680711746216, | |
| "learning_rate": 4.21455938697318e-05, | |
| "loss": 0.2947, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 21.75862068965517, | |
| "grad_norm": 0.8190564513206482, | |
| "learning_rate": 4.203612479474548e-05, | |
| "loss": 0.1653, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 21.79310344827586, | |
| "grad_norm": 4.2845635414123535, | |
| "learning_rate": 4.1926655719759176e-05, | |
| "loss": 0.1636, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 21.82758620689655, | |
| "grad_norm": 4.332767009735107, | |
| "learning_rate": 4.1817186644772855e-05, | |
| "loss": 0.2388, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 21.862068965517242, | |
| "grad_norm": 8.83471393585205, | |
| "learning_rate": 4.1707717569786535e-05, | |
| "loss": 0.3177, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 21.896551724137932, | |
| "grad_norm": 0.2615772485733032, | |
| "learning_rate": 4.159824849480022e-05, | |
| "loss": 0.3468, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 21.93103448275862, | |
| "grad_norm": 1.6727783679962158, | |
| "learning_rate": 4.148877941981391e-05, | |
| "loss": 0.5736, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 21.96551724137931, | |
| "grad_norm": 0.6974768042564392, | |
| "learning_rate": 4.1379310344827587e-05, | |
| "loss": 0.1321, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "grad_norm": 0.36256420612335205, | |
| "learning_rate": 4.126984126984127e-05, | |
| "loss": 0.0793, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.569184754221156, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6209601528534142, | |
| "eval_loss": 1.3289726972579956, | |
| "eval_precision_macro": 0.5959995620321707, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6358640227526097, | |
| "eval_recall_macro": 0.5650793650793651, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 1.7261, | |
| "eval_samples_per_second": 76.472, | |
| "eval_steps_per_second": 9.849, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 22.03448275862069, | |
| "grad_norm": 1.9399197101593018, | |
| "learning_rate": 4.116037219485496e-05, | |
| "loss": 0.354, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 22.06896551724138, | |
| "grad_norm": 0.5011938810348511, | |
| "learning_rate": 4.105090311986864e-05, | |
| "loss": 0.0993, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 22.103448275862068, | |
| "grad_norm": 0.2772790193557739, | |
| "learning_rate": 4.094143404488232e-05, | |
| "loss": 0.2445, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 22.137931034482758, | |
| "grad_norm": 11.734522819519043, | |
| "learning_rate": 4.083196496989601e-05, | |
| "loss": 0.5215, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 22.17241379310345, | |
| "grad_norm": 5.657594680786133, | |
| "learning_rate": 4.072249589490969e-05, | |
| "loss": 0.2948, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 22.20689655172414, | |
| "grad_norm": 5.29397439956665, | |
| "learning_rate": 4.061302681992337e-05, | |
| "loss": 0.367, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 22.24137931034483, | |
| "grad_norm": 1.1745647192001343, | |
| "learning_rate": 4.050355774493706e-05, | |
| "loss": 0.1183, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 22.275862068965516, | |
| "grad_norm": 10.664090156555176, | |
| "learning_rate": 4.039408866995074e-05, | |
| "loss": 0.1827, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 22.310344827586206, | |
| "grad_norm": 2.1371960639953613, | |
| "learning_rate": 4.028461959496442e-05, | |
| "loss": 0.1411, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 22.344827586206897, | |
| "grad_norm": 14.131455421447754, | |
| "learning_rate": 4.017515051997811e-05, | |
| "loss": 0.487, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 22.379310344827587, | |
| "grad_norm": 0.5914010405540466, | |
| "learning_rate": 4.0065681444991794e-05, | |
| "loss": 0.0997, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 22.413793103448278, | |
| "grad_norm": 8.857903480529785, | |
| "learning_rate": 3.9956212370005474e-05, | |
| "loss": 0.4644, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 22.448275862068964, | |
| "grad_norm": 1.2613619565963745, | |
| "learning_rate": 3.984674329501916e-05, | |
| "loss": 0.0965, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 22.482758620689655, | |
| "grad_norm": 4.910549640655518, | |
| "learning_rate": 3.9737274220032846e-05, | |
| "loss": 0.6465, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 22.517241379310345, | |
| "grad_norm": 1.887235164642334, | |
| "learning_rate": 3.9627805145046526e-05, | |
| "loss": 0.1883, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 22.551724137931036, | |
| "grad_norm": 5.57004451751709, | |
| "learning_rate": 3.9518336070060205e-05, | |
| "loss": 0.2007, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 22.586206896551722, | |
| "grad_norm": 7.169775485992432, | |
| "learning_rate": 3.94088669950739e-05, | |
| "loss": 0.1982, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 22.620689655172413, | |
| "grad_norm": 0.8967238068580627, | |
| "learning_rate": 3.929939792008758e-05, | |
| "loss": 0.1722, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 22.655172413793103, | |
| "grad_norm": 3.6087558269500732, | |
| "learning_rate": 3.918992884510126e-05, | |
| "loss": 0.2057, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 22.689655172413794, | |
| "grad_norm": 2.1949057579040527, | |
| "learning_rate": 3.908045977011495e-05, | |
| "loss": 0.5147, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 22.724137931034484, | |
| "grad_norm": 0.6924113631248474, | |
| "learning_rate": 3.897099069512863e-05, | |
| "loss": 0.2777, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 22.75862068965517, | |
| "grad_norm": 0.32273250818252563, | |
| "learning_rate": 3.886152162014231e-05, | |
| "loss": 0.0778, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 22.79310344827586, | |
| "grad_norm": 3.3658881187438965, | |
| "learning_rate": 3.8752052545155995e-05, | |
| "loss": 0.1048, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 22.82758620689655, | |
| "grad_norm": 6.8869194984436035, | |
| "learning_rate": 3.864258347016968e-05, | |
| "loss": 0.2943, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 22.862068965517242, | |
| "grad_norm": 3.4153761863708496, | |
| "learning_rate": 3.853311439518336e-05, | |
| "loss": 0.5214, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 22.896551724137932, | |
| "grad_norm": 22.111948013305664, | |
| "learning_rate": 3.842364532019704e-05, | |
| "loss": 0.1767, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 22.93103448275862, | |
| "grad_norm": 1.5236107110977173, | |
| "learning_rate": 3.831417624521073e-05, | |
| "loss": 0.0809, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 22.96551724137931, | |
| "grad_norm": 8.651114463806152, | |
| "learning_rate": 3.820470717022441e-05, | |
| "loss": 0.3069, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "grad_norm": 2.85559344291687, | |
| "learning_rate": 3.809523809523809e-05, | |
| "loss": 0.1761, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_accuracy": 0.6212121212121212, | |
| "eval_f1_macro": 0.5572308233638282, | |
| "eval_f1_micro": 0.6212121212121212, | |
| "eval_f1_weighted": 0.6032445763049211, | |
| "eval_loss": 1.4283698797225952, | |
| "eval_precision_macro": 0.6453514739229025, | |
| "eval_precision_micro": 0.6212121212121212, | |
| "eval_precision_weighted": 0.6562950937950938, | |
| "eval_recall_macro": 0.5515646258503402, | |
| "eval_recall_micro": 0.6212121212121212, | |
| "eval_recall_weighted": 0.6212121212121212, | |
| "eval_runtime": 1.7048, | |
| "eval_samples_per_second": 77.427, | |
| "eval_steps_per_second": 9.972, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 23.03448275862069, | |
| "grad_norm": 10.627945899963379, | |
| "learning_rate": 3.7985769020251785e-05, | |
| "loss": 0.2271, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 23.06896551724138, | |
| "grad_norm": 2.46520733833313, | |
| "learning_rate": 3.7876299945265465e-05, | |
| "loss": 0.0865, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 23.103448275862068, | |
| "grad_norm": 14.325843811035156, | |
| "learning_rate": 3.7766830870279144e-05, | |
| "loss": 0.5699, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 23.137931034482758, | |
| "grad_norm": 3.0035719871520996, | |
| "learning_rate": 3.765736179529283e-05, | |
| "loss": 0.2166, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 23.17241379310345, | |
| "grad_norm": 7.00339412689209, | |
| "learning_rate": 3.7547892720306517e-05, | |
| "loss": 0.3422, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 23.20689655172414, | |
| "grad_norm": 4.217713356018066, | |
| "learning_rate": 3.7438423645320196e-05, | |
| "loss": 0.1615, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 23.24137931034483, | |
| "grad_norm": 3.0487077236175537, | |
| "learning_rate": 3.732895457033388e-05, | |
| "loss": 0.3146, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 23.275862068965516, | |
| "grad_norm": 11.421404838562012, | |
| "learning_rate": 3.721948549534757e-05, | |
| "loss": 0.2691, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 23.310344827586206, | |
| "grad_norm": 0.3206021189689636, | |
| "learning_rate": 3.711001642036125e-05, | |
| "loss": 0.0665, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 23.344827586206897, | |
| "grad_norm": 1.73373544216156, | |
| "learning_rate": 3.700054734537493e-05, | |
| "loss": 0.154, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 23.379310344827587, | |
| "grad_norm": 0.29473578929901123, | |
| "learning_rate": 3.689107827038862e-05, | |
| "loss": 0.344, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 23.413793103448278, | |
| "grad_norm": 3.3459489345550537, | |
| "learning_rate": 3.67816091954023e-05, | |
| "loss": 0.1159, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 23.448275862068964, | |
| "grad_norm": 1.0497227907180786, | |
| "learning_rate": 3.667214012041598e-05, | |
| "loss": 0.093, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 23.482758620689655, | |
| "grad_norm": 9.836808204650879, | |
| "learning_rate": 3.656267104542967e-05, | |
| "loss": 0.1138, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 23.517241379310345, | |
| "grad_norm": 0.22488915920257568, | |
| "learning_rate": 3.645320197044335e-05, | |
| "loss": 0.0985, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 23.551724137931036, | |
| "grad_norm": 2.9141793251037598, | |
| "learning_rate": 3.634373289545703e-05, | |
| "loss": 0.1183, | |
| "step": 1366 | |
| }, | |
| { | |
| "epoch": 23.586206896551722, | |
| "grad_norm": 7.301143646240234, | |
| "learning_rate": 3.623426382047072e-05, | |
| "loss": 0.1026, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 23.620689655172413, | |
| "grad_norm": 2.478011131286621, | |
| "learning_rate": 3.6124794745484404e-05, | |
| "loss": 0.175, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 23.655172413793103, | |
| "grad_norm": 0.43685877323150635, | |
| "learning_rate": 3.601532567049808e-05, | |
| "loss": 0.0807, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 23.689655172413794, | |
| "grad_norm": 0.7601661086082458, | |
| "learning_rate": 3.590585659551177e-05, | |
| "loss": 0.2639, | |
| "step": 1374 | |
| }, | |
| { | |
| "epoch": 23.724137931034484, | |
| "grad_norm": 3.5471153259277344, | |
| "learning_rate": 3.5796387520525456e-05, | |
| "loss": 0.1622, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 23.75862068965517, | |
| "grad_norm": 5.943011283874512, | |
| "learning_rate": 3.5686918445539135e-05, | |
| "loss": 0.2493, | |
| "step": 1378 | |
| }, | |
| { | |
| "epoch": 23.79310344827586, | |
| "grad_norm": 0.5235481858253479, | |
| "learning_rate": 3.5577449370552815e-05, | |
| "loss": 0.0738, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 23.82758620689655, | |
| "grad_norm": 0.9073754549026489, | |
| "learning_rate": 3.546798029556651e-05, | |
| "loss": 0.2598, | |
| "step": 1382 | |
| }, | |
| { | |
| "epoch": 23.862068965517242, | |
| "grad_norm": 0.35284557938575745, | |
| "learning_rate": 3.535851122058019e-05, | |
| "loss": 0.0684, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 23.896551724137932, | |
| "grad_norm": 7.837213039398193, | |
| "learning_rate": 3.5249042145593867e-05, | |
| "loss": 0.1072, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 23.93103448275862, | |
| "grad_norm": 10.070199966430664, | |
| "learning_rate": 3.513957307060756e-05, | |
| "loss": 0.2412, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 23.96551724137931, | |
| "grad_norm": 22.912311553955078, | |
| "learning_rate": 3.503010399562124e-05, | |
| "loss": 0.379, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "grad_norm": 11.414230346679688, | |
| "learning_rate": 3.492063492063492e-05, | |
| "loss": 0.1714, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5781647222832837, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6343655897181453, | |
| "eval_loss": 1.2994105815887451, | |
| "eval_precision_macro": 0.5899192043967345, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6461095703103377, | |
| "eval_recall_macro": 0.5727739984882841, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 1.7262, | |
| "eval_samples_per_second": 76.469, | |
| "eval_steps_per_second": 9.848, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 24.03448275862069, | |
| "grad_norm": 1.8347530364990234, | |
| "learning_rate": 3.4811165845648605e-05, | |
| "loss": 0.0856, | |
| "step": 1394 | |
| }, | |
| { | |
| "epoch": 24.06896551724138, | |
| "grad_norm": 0.23063726723194122, | |
| "learning_rate": 3.470169677066229e-05, | |
| "loss": 0.0486, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 24.103448275862068, | |
| "grad_norm": 13.704474449157715, | |
| "learning_rate": 3.459222769567597e-05, | |
| "loss": 0.0688, | |
| "step": 1398 | |
| }, | |
| { | |
| "epoch": 24.137931034482758, | |
| "grad_norm": 0.7486428022384644, | |
| "learning_rate": 3.4482758620689657e-05, | |
| "loss": 0.0726, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 24.17241379310345, | |
| "grad_norm": 5.045259952545166, | |
| "learning_rate": 3.437328954570334e-05, | |
| "loss": 0.5948, | |
| "step": 1402 | |
| }, | |
| { | |
| "epoch": 24.20689655172414, | |
| "grad_norm": 17.36922836303711, | |
| "learning_rate": 3.426382047071702e-05, | |
| "loss": 0.1763, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 24.24137931034483, | |
| "grad_norm": 4.404787063598633, | |
| "learning_rate": 3.415435139573071e-05, | |
| "loss": 0.0887, | |
| "step": 1406 | |
| }, | |
| { | |
| "epoch": 24.275862068965516, | |
| "grad_norm": 1.7757786512374878, | |
| "learning_rate": 3.4044882320744395e-05, | |
| "loss": 0.0816, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 24.310344827586206, | |
| "grad_norm": 10.455934524536133, | |
| "learning_rate": 3.3935413245758074e-05, | |
| "loss": 0.3544, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 24.344827586206897, | |
| "grad_norm": 1.8070427179336548, | |
| "learning_rate": 3.3825944170771754e-05, | |
| "loss": 0.3629, | |
| "step": 1412 | |
| }, | |
| { | |
| "epoch": 24.379310344827587, | |
| "grad_norm": 12.575775146484375, | |
| "learning_rate": 3.371647509578545e-05, | |
| "loss": 0.1447, | |
| "step": 1414 | |
| }, | |
| { | |
| "epoch": 24.413793103448278, | |
| "grad_norm": 17.860628128051758, | |
| "learning_rate": 3.3607006020799126e-05, | |
| "loss": 0.2491, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 24.448275862068964, | |
| "grad_norm": 0.40125876665115356, | |
| "learning_rate": 3.3497536945812806e-05, | |
| "loss": 0.0505, | |
| "step": 1418 | |
| }, | |
| { | |
| "epoch": 24.482758620689655, | |
| "grad_norm": 0.3735012710094452, | |
| "learning_rate": 3.338806787082649e-05, | |
| "loss": 0.0842, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 24.517241379310345, | |
| "grad_norm": 10.689638137817383, | |
| "learning_rate": 3.327859879584018e-05, | |
| "loss": 0.4145, | |
| "step": 1422 | |
| }, | |
| { | |
| "epoch": 24.551724137931036, | |
| "grad_norm": 0.5746628642082214, | |
| "learning_rate": 3.316912972085386e-05, | |
| "loss": 0.1959, | |
| "step": 1424 | |
| }, | |
| { | |
| "epoch": 24.586206896551722, | |
| "grad_norm": 1.2556354999542236, | |
| "learning_rate": 3.3059660645867544e-05, | |
| "loss": 0.1955, | |
| "step": 1426 | |
| }, | |
| { | |
| "epoch": 24.620689655172413, | |
| "grad_norm": 0.3656192421913147, | |
| "learning_rate": 3.295019157088123e-05, | |
| "loss": 0.0546, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 24.655172413793103, | |
| "grad_norm": 0.373201847076416, | |
| "learning_rate": 3.284072249589491e-05, | |
| "loss": 0.0566, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 24.689655172413794, | |
| "grad_norm": 10.170561790466309, | |
| "learning_rate": 3.2731253420908596e-05, | |
| "loss": 0.129, | |
| "step": 1432 | |
| }, | |
| { | |
| "epoch": 24.724137931034484, | |
| "grad_norm": 0.8907169103622437, | |
| "learning_rate": 3.262178434592228e-05, | |
| "loss": 0.0553, | |
| "step": 1434 | |
| }, | |
| { | |
| "epoch": 24.75862068965517, | |
| "grad_norm": 9.43267822265625, | |
| "learning_rate": 3.251231527093596e-05, | |
| "loss": 0.149, | |
| "step": 1436 | |
| }, | |
| { | |
| "epoch": 24.79310344827586, | |
| "grad_norm": 8.570932388305664, | |
| "learning_rate": 3.240284619594964e-05, | |
| "loss": 0.1292, | |
| "step": 1438 | |
| }, | |
| { | |
| "epoch": 24.82758620689655, | |
| "grad_norm": 13.18994426727295, | |
| "learning_rate": 3.2293377120963334e-05, | |
| "loss": 0.2692, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 24.862068965517242, | |
| "grad_norm": 0.47455695271492004, | |
| "learning_rate": 3.218390804597701e-05, | |
| "loss": 0.41, | |
| "step": 1442 | |
| }, | |
| { | |
| "epoch": 24.896551724137932, | |
| "grad_norm": 1.234464406967163, | |
| "learning_rate": 3.207443897099069e-05, | |
| "loss": 0.056, | |
| "step": 1444 | |
| }, | |
| { | |
| "epoch": 24.93103448275862, | |
| "grad_norm": 1.0157136917114258, | |
| "learning_rate": 3.196496989600438e-05, | |
| "loss": 0.1239, | |
| "step": 1446 | |
| }, | |
| { | |
| "epoch": 24.96551724137931, | |
| "grad_norm": 0.6129853129386902, | |
| "learning_rate": 3.1855500821018065e-05, | |
| "loss": 0.0641, | |
| "step": 1448 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "grad_norm": 8.584732055664062, | |
| "learning_rate": 3.1746031746031745e-05, | |
| "loss": 0.465, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_accuracy": 0.6136363636363636, | |
| "eval_f1_macro": 0.558068783068783, | |
| "eval_f1_micro": 0.6136363636363636, | |
| "eval_f1_weighted": 0.6134259259259258, | |
| "eval_loss": 1.4011425971984863, | |
| "eval_precision_macro": 0.5662329205186348, | |
| "eval_precision_micro": 0.6136363636363636, | |
| "eval_precision_weighted": 0.6187895437895439, | |
| "eval_recall_macro": 0.5556084656084657, | |
| "eval_recall_micro": 0.6136363636363636, | |
| "eval_recall_weighted": 0.6136363636363636, | |
| "eval_runtime": 1.7373, | |
| "eval_samples_per_second": 75.98, | |
| "eval_steps_per_second": 9.785, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 25.03448275862069, | |
| "grad_norm": 0.2674843966960907, | |
| "learning_rate": 3.163656267104543e-05, | |
| "loss": 0.063, | |
| "step": 1452 | |
| }, | |
| { | |
| "epoch": 25.06896551724138, | |
| "grad_norm": 0.20413754880428314, | |
| "learning_rate": 3.152709359605912e-05, | |
| "loss": 0.2553, | |
| "step": 1454 | |
| }, | |
| { | |
| "epoch": 25.103448275862068, | |
| "grad_norm": 1.4728457927703857, | |
| "learning_rate": 3.1417624521072797e-05, | |
| "loss": 0.0623, | |
| "step": 1456 | |
| }, | |
| { | |
| "epoch": 25.137931034482758, | |
| "grad_norm": 1.164231300354004, | |
| "learning_rate": 3.130815544608648e-05, | |
| "loss": 0.0657, | |
| "step": 1458 | |
| }, | |
| { | |
| "epoch": 25.17241379310345, | |
| "grad_norm": 0.4934793710708618, | |
| "learning_rate": 3.119868637110017e-05, | |
| "loss": 0.1067, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 25.20689655172414, | |
| "grad_norm": 11.123279571533203, | |
| "learning_rate": 3.108921729611385e-05, | |
| "loss": 0.5103, | |
| "step": 1462 | |
| }, | |
| { | |
| "epoch": 25.24137931034483, | |
| "grad_norm": 0.14121675491333008, | |
| "learning_rate": 3.097974822112753e-05, | |
| "loss": 0.0531, | |
| "step": 1464 | |
| }, | |
| { | |
| "epoch": 25.275862068965516, | |
| "grad_norm": 0.2784970998764038, | |
| "learning_rate": 3.087027914614122e-05, | |
| "loss": 0.0509, | |
| "step": 1466 | |
| }, | |
| { | |
| "epoch": 25.310344827586206, | |
| "grad_norm": 2.157268762588501, | |
| "learning_rate": 3.07608100711549e-05, | |
| "loss": 0.0852, | |
| "step": 1468 | |
| }, | |
| { | |
| "epoch": 25.344827586206897, | |
| "grad_norm": 2.208381414413452, | |
| "learning_rate": 3.065134099616858e-05, | |
| "loss": 0.0847, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 25.379310344827587, | |
| "grad_norm": 0.3056719899177551, | |
| "learning_rate": 3.0541871921182266e-05, | |
| "loss": 0.044, | |
| "step": 1472 | |
| }, | |
| { | |
| "epoch": 25.413793103448278, | |
| "grad_norm": 0.2943095266819, | |
| "learning_rate": 3.0432402846195952e-05, | |
| "loss": 0.1533, | |
| "step": 1474 | |
| }, | |
| { | |
| "epoch": 25.448275862068964, | |
| "grad_norm": 7.445497035980225, | |
| "learning_rate": 3.0322933771209632e-05, | |
| "loss": 0.254, | |
| "step": 1476 | |
| }, | |
| { | |
| "epoch": 25.482758620689655, | |
| "grad_norm": 1.454938292503357, | |
| "learning_rate": 3.021346469622332e-05, | |
| "loss": 0.0573, | |
| "step": 1478 | |
| }, | |
| { | |
| "epoch": 25.517241379310345, | |
| "grad_norm": 0.25220704078674316, | |
| "learning_rate": 3.0103995621237e-05, | |
| "loss": 0.0506, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 25.551724137931036, | |
| "grad_norm": 1.2520936727523804, | |
| "learning_rate": 2.9994526546250684e-05, | |
| "loss": 0.0533, | |
| "step": 1482 | |
| }, | |
| { | |
| "epoch": 25.586206896551722, | |
| "grad_norm": 0.5618668794631958, | |
| "learning_rate": 2.988505747126437e-05, | |
| "loss": 0.05, | |
| "step": 1484 | |
| }, | |
| { | |
| "epoch": 25.620689655172413, | |
| "grad_norm": 6.944793701171875, | |
| "learning_rate": 2.9775588396278053e-05, | |
| "loss": 0.1004, | |
| "step": 1486 | |
| }, | |
| { | |
| "epoch": 25.655172413793103, | |
| "grad_norm": 0.36061984300613403, | |
| "learning_rate": 2.9666119321291736e-05, | |
| "loss": 0.0442, | |
| "step": 1488 | |
| }, | |
| { | |
| "epoch": 25.689655172413794, | |
| "grad_norm": 13.36592960357666, | |
| "learning_rate": 2.9556650246305422e-05, | |
| "loss": 0.1224, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 25.724137931034484, | |
| "grad_norm": 1.499154806137085, | |
| "learning_rate": 2.9447181171319105e-05, | |
| "loss": 0.0552, | |
| "step": 1492 | |
| }, | |
| { | |
| "epoch": 25.75862068965517, | |
| "grad_norm": 1.4718117713928223, | |
| "learning_rate": 2.9337712096332788e-05, | |
| "loss": 0.1021, | |
| "step": 1494 | |
| }, | |
| { | |
| "epoch": 25.79310344827586, | |
| "grad_norm": 25.769798278808594, | |
| "learning_rate": 2.9228243021346467e-05, | |
| "loss": 0.6279, | |
| "step": 1496 | |
| }, | |
| { | |
| "epoch": 25.82758620689655, | |
| "grad_norm": 1.972213625907898, | |
| "learning_rate": 2.9118773946360157e-05, | |
| "loss": 0.2089, | |
| "step": 1498 | |
| }, | |
| { | |
| "epoch": 25.862068965517242, | |
| "grad_norm": 0.9450302124023438, | |
| "learning_rate": 2.900930487137384e-05, | |
| "loss": 0.0474, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 25.896551724137932, | |
| "grad_norm": 2.7113711833953857, | |
| "learning_rate": 2.889983579638752e-05, | |
| "loss": 0.4025, | |
| "step": 1502 | |
| }, | |
| { | |
| "epoch": 25.93103448275862, | |
| "grad_norm": 23.863096237182617, | |
| "learning_rate": 2.879036672140121e-05, | |
| "loss": 0.3943, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 25.96551724137931, | |
| "grad_norm": 0.5137322545051575, | |
| "learning_rate": 2.8680897646414888e-05, | |
| "loss": 0.0574, | |
| "step": 1506 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "grad_norm": 9.085604667663574, | |
| "learning_rate": 2.857142857142857e-05, | |
| "loss": 0.2203, | |
| "step": 1508 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5740664325639695, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6265666159279011, | |
| "eval_loss": 1.4700660705566406, | |
| "eval_precision_macro": 0.6167028006681489, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6552921047278235, | |
| "eval_recall_macro": 0.5675888133030991, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 1.7235, | |
| "eval_samples_per_second": 76.588, | |
| "eval_steps_per_second": 9.864, | |
| "step": 1508 | |
| }, | |
| { | |
| "epoch": 26.03448275862069, | |
| "grad_norm": 4.422103404998779, | |
| "learning_rate": 2.8461959496442257e-05, | |
| "loss": 0.0811, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 26.06896551724138, | |
| "grad_norm": 4.039704322814941, | |
| "learning_rate": 2.835249042145594e-05, | |
| "loss": 0.3148, | |
| "step": 1512 | |
| }, | |
| { | |
| "epoch": 26.103448275862068, | |
| "grad_norm": 0.1570342779159546, | |
| "learning_rate": 2.8243021346469623e-05, | |
| "loss": 0.0437, | |
| "step": 1514 | |
| }, | |
| { | |
| "epoch": 26.137931034482758, | |
| "grad_norm": 6.058873176574707, | |
| "learning_rate": 2.813355227148331e-05, | |
| "loss": 0.1608, | |
| "step": 1516 | |
| }, | |
| { | |
| "epoch": 26.17241379310345, | |
| "grad_norm": 3.495460033416748, | |
| "learning_rate": 2.8024083196496992e-05, | |
| "loss": 0.0555, | |
| "step": 1518 | |
| }, | |
| { | |
| "epoch": 26.20689655172414, | |
| "grad_norm": 7.322564601898193, | |
| "learning_rate": 2.7914614121510675e-05, | |
| "loss": 0.0886, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 26.24137931034483, | |
| "grad_norm": 0.18538200855255127, | |
| "learning_rate": 2.7805145046524354e-05, | |
| "loss": 0.0562, | |
| "step": 1522 | |
| }, | |
| { | |
| "epoch": 26.275862068965516, | |
| "grad_norm": 0.3719606101512909, | |
| "learning_rate": 2.7695675971538044e-05, | |
| "loss": 0.041, | |
| "step": 1524 | |
| }, | |
| { | |
| "epoch": 26.310344827586206, | |
| "grad_norm": 0.2550795376300812, | |
| "learning_rate": 2.7586206896551727e-05, | |
| "loss": 0.1232, | |
| "step": 1526 | |
| }, | |
| { | |
| "epoch": 26.344827586206897, | |
| "grad_norm": 0.3839069604873657, | |
| "learning_rate": 2.7476737821565406e-05, | |
| "loss": 0.204, | |
| "step": 1528 | |
| }, | |
| { | |
| "epoch": 26.379310344827587, | |
| "grad_norm": 13.937886238098145, | |
| "learning_rate": 2.7367268746579096e-05, | |
| "loss": 0.4575, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 26.413793103448278, | |
| "grad_norm": 3.547407865524292, | |
| "learning_rate": 2.7257799671592775e-05, | |
| "loss": 0.2855, | |
| "step": 1532 | |
| }, | |
| { | |
| "epoch": 26.448275862068964, | |
| "grad_norm": 18.570228576660156, | |
| "learning_rate": 2.7148330596606458e-05, | |
| "loss": 0.1721, | |
| "step": 1534 | |
| }, | |
| { | |
| "epoch": 26.482758620689655, | |
| "grad_norm": 0.25473734736442566, | |
| "learning_rate": 2.7038861521620144e-05, | |
| "loss": 0.0521, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 26.517241379310345, | |
| "grad_norm": 3.582709312438965, | |
| "learning_rate": 2.6929392446633827e-05, | |
| "loss": 0.0713, | |
| "step": 1538 | |
| }, | |
| { | |
| "epoch": 26.551724137931036, | |
| "grad_norm": 0.18364673852920532, | |
| "learning_rate": 2.681992337164751e-05, | |
| "loss": 0.219, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 26.586206896551722, | |
| "grad_norm": 0.24068348109722137, | |
| "learning_rate": 2.6710454296661196e-05, | |
| "loss": 0.3014, | |
| "step": 1542 | |
| }, | |
| { | |
| "epoch": 26.620689655172413, | |
| "grad_norm": 0.4433784782886505, | |
| "learning_rate": 2.660098522167488e-05, | |
| "loss": 0.0689, | |
| "step": 1544 | |
| }, | |
| { | |
| "epoch": 26.655172413793103, | |
| "grad_norm": 0.49119868874549866, | |
| "learning_rate": 2.6491516146688562e-05, | |
| "loss": 0.0461, | |
| "step": 1546 | |
| }, | |
| { | |
| "epoch": 26.689655172413794, | |
| "grad_norm": 5.021825313568115, | |
| "learning_rate": 2.638204707170224e-05, | |
| "loss": 0.2385, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 26.724137931034484, | |
| "grad_norm": 0.26685020327568054, | |
| "learning_rate": 2.627257799671593e-05, | |
| "loss": 0.0427, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 26.75862068965517, | |
| "grad_norm": 1.1268709897994995, | |
| "learning_rate": 2.6163108921729614e-05, | |
| "loss": 0.0586, | |
| "step": 1552 | |
| }, | |
| { | |
| "epoch": 26.79310344827586, | |
| "grad_norm": 0.20585598051548004, | |
| "learning_rate": 2.6053639846743293e-05, | |
| "loss": 0.1416, | |
| "step": 1554 | |
| }, | |
| { | |
| "epoch": 26.82758620689655, | |
| "grad_norm": 0.39748844504356384, | |
| "learning_rate": 2.5944170771756983e-05, | |
| "loss": 0.0484, | |
| "step": 1556 | |
| }, | |
| { | |
| "epoch": 26.862068965517242, | |
| "grad_norm": 0.6568892598152161, | |
| "learning_rate": 2.5834701696770662e-05, | |
| "loss": 0.0427, | |
| "step": 1558 | |
| }, | |
| { | |
| "epoch": 26.896551724137932, | |
| "grad_norm": 0.6816446185112, | |
| "learning_rate": 2.5725232621784345e-05, | |
| "loss": 0.3039, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 26.93103448275862, | |
| "grad_norm": 0.5687500238418579, | |
| "learning_rate": 2.561576354679803e-05, | |
| "loss": 0.0555, | |
| "step": 1562 | |
| }, | |
| { | |
| "epoch": 26.96551724137931, | |
| "grad_norm": 0.29769548773765564, | |
| "learning_rate": 2.5506294471811714e-05, | |
| "loss": 0.3812, | |
| "step": 1564 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "grad_norm": 0.5284446477890015, | |
| "learning_rate": 2.5396825396825397e-05, | |
| "loss": 0.0574, | |
| "step": 1566 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_accuracy": 0.6363636363636364, | |
| "eval_f1_macro": 0.5799920937164608, | |
| "eval_f1_micro": 0.6363636363636364, | |
| "eval_f1_weighted": 0.6352104119661408, | |
| "eval_loss": 1.451111078262329, | |
| "eval_precision_macro": 0.6073050209589084, | |
| "eval_precision_micro": 0.6363636363636364, | |
| "eval_precision_weighted": 0.65455105239588, | |
| "eval_recall_macro": 0.5738397581254724, | |
| "eval_recall_micro": 0.6363636363636364, | |
| "eval_recall_weighted": 0.6363636363636364, | |
| "eval_runtime": 1.6969, | |
| "eval_samples_per_second": 77.788, | |
| "eval_steps_per_second": 10.018, | |
| "step": 1566 | |
| }, | |
| { | |
| "epoch": 27.03448275862069, | |
| "grad_norm": 5.421750068664551, | |
| "learning_rate": 2.5287356321839083e-05, | |
| "loss": 0.0798, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 27.06896551724138, | |
| "grad_norm": 0.11500166356563568, | |
| "learning_rate": 2.5177887246852766e-05, | |
| "loss": 0.0314, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 27.103448275862068, | |
| "grad_norm": 0.19810307025909424, | |
| "learning_rate": 2.506841817186645e-05, | |
| "loss": 0.2181, | |
| "step": 1572 | |
| }, | |
| { | |
| "epoch": 27.137931034482758, | |
| "grad_norm": 0.21842528879642487, | |
| "learning_rate": 2.4958949096880132e-05, | |
| "loss": 0.041, | |
| "step": 1574 | |
| }, | |
| { | |
| "epoch": 27.17241379310345, | |
| "grad_norm": 3.9452264308929443, | |
| "learning_rate": 2.4849480021893818e-05, | |
| "loss": 0.0789, | |
| "step": 1576 | |
| }, | |
| { | |
| "epoch": 27.20689655172414, | |
| "grad_norm": 0.18447476625442505, | |
| "learning_rate": 2.4740010946907498e-05, | |
| "loss": 0.0376, | |
| "step": 1578 | |
| }, | |
| { | |
| "epoch": 27.24137931034483, | |
| "grad_norm": 1.6792452335357666, | |
| "learning_rate": 2.4630541871921184e-05, | |
| "loss": 0.049, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 27.275862068965516, | |
| "grad_norm": 0.5313147306442261, | |
| "learning_rate": 2.4521072796934867e-05, | |
| "loss": 0.0555, | |
| "step": 1582 | |
| }, | |
| { | |
| "epoch": 27.310344827586206, | |
| "grad_norm": 0.2032177746295929, | |
| "learning_rate": 2.441160372194855e-05, | |
| "loss": 0.0846, | |
| "step": 1584 | |
| }, | |
| { | |
| "epoch": 27.344827586206897, | |
| "grad_norm": 5.808465480804443, | |
| "learning_rate": 2.4302134646962236e-05, | |
| "loss": 0.308, | |
| "step": 1586 | |
| }, | |
| { | |
| "epoch": 27.379310344827587, | |
| "grad_norm": 15.264229774475098, | |
| "learning_rate": 2.419266557197592e-05, | |
| "loss": 0.262, | |
| "step": 1588 | |
| }, | |
| { | |
| "epoch": 27.413793103448278, | |
| "grad_norm": 0.4792754650115967, | |
| "learning_rate": 2.40831964969896e-05, | |
| "loss": 0.2704, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 27.448275862068964, | |
| "grad_norm": 1.3727836608886719, | |
| "learning_rate": 2.3973727422003288e-05, | |
| "loss": 0.2656, | |
| "step": 1592 | |
| }, | |
| { | |
| "epoch": 27.482758620689655, | |
| "grad_norm": 0.4460311532020569, | |
| "learning_rate": 2.3864258347016967e-05, | |
| "loss": 0.0396, | |
| "step": 1594 | |
| }, | |
| { | |
| "epoch": 27.517241379310345, | |
| "grad_norm": 0.4063487946987152, | |
| "learning_rate": 2.3754789272030653e-05, | |
| "loss": 0.074, | |
| "step": 1596 | |
| }, | |
| { | |
| "epoch": 27.551724137931036, | |
| "grad_norm": 0.8473443984985352, | |
| "learning_rate": 2.3645320197044336e-05, | |
| "loss": 0.0509, | |
| "step": 1598 | |
| }, | |
| { | |
| "epoch": 27.586206896551722, | |
| "grad_norm": 0.139051616191864, | |
| "learning_rate": 2.353585112205802e-05, | |
| "loss": 0.0326, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 27.620689655172413, | |
| "grad_norm": 8.467299461364746, | |
| "learning_rate": 2.3426382047071705e-05, | |
| "loss": 0.0749, | |
| "step": 1602 | |
| }, | |
| { | |
| "epoch": 27.655172413793103, | |
| "grad_norm": 0.3194819688796997, | |
| "learning_rate": 2.3316912972085385e-05, | |
| "loss": 0.0845, | |
| "step": 1604 | |
| }, | |
| { | |
| "epoch": 27.689655172413794, | |
| "grad_norm": 0.4060407280921936, | |
| "learning_rate": 2.320744389709907e-05, | |
| "loss": 0.0819, | |
| "step": 1606 | |
| }, | |
| { | |
| "epoch": 27.724137931034484, | |
| "grad_norm": 0.14499817788600922, | |
| "learning_rate": 2.3097974822112754e-05, | |
| "loss": 0.0405, | |
| "step": 1608 | |
| }, | |
| { | |
| "epoch": 27.75862068965517, | |
| "grad_norm": 0.17447809875011444, | |
| "learning_rate": 2.2988505747126437e-05, | |
| "loss": 0.0356, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 27.79310344827586, | |
| "grad_norm": 3.4749953746795654, | |
| "learning_rate": 2.2879036672140123e-05, | |
| "loss": 0.0674, | |
| "step": 1612 | |
| }, | |
| { | |
| "epoch": 27.82758620689655, | |
| "grad_norm": 0.12591542303562164, | |
| "learning_rate": 2.2769567597153806e-05, | |
| "loss": 0.1736, | |
| "step": 1614 | |
| }, | |
| { | |
| "epoch": 27.862068965517242, | |
| "grad_norm": 0.7413695454597473, | |
| "learning_rate": 2.266009852216749e-05, | |
| "loss": 0.044, | |
| "step": 1616 | |
| }, | |
| { | |
| "epoch": 27.896551724137932, | |
| "grad_norm": 0.27023714780807495, | |
| "learning_rate": 2.2550629447181175e-05, | |
| "loss": 0.0764, | |
| "step": 1618 | |
| }, | |
| { | |
| "epoch": 27.93103448275862, | |
| "grad_norm": 0.7378279566764832, | |
| "learning_rate": 2.2441160372194854e-05, | |
| "loss": 0.1553, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 27.96551724137931, | |
| "grad_norm": 12.137617111206055, | |
| "learning_rate": 2.233169129720854e-05, | |
| "loss": 0.5215, | |
| "step": 1622 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "grad_norm": 0.11637034267187119, | |
| "learning_rate": 2.2222222222222223e-05, | |
| "loss": 0.0399, | |
| "step": 1624 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_accuracy": 0.6060606060606061, | |
| "eval_f1_macro": 0.5674308896409629, | |
| "eval_f1_micro": 0.6060606060606061, | |
| "eval_f1_weighted": 0.6133032153841548, | |
| "eval_loss": 1.4920562505722046, | |
| "eval_precision_macro": 0.5933232665951561, | |
| "eval_precision_micro": 0.6060606060606061, | |
| "eval_precision_weighted": 0.6389820020318553, | |
| "eval_recall_macro": 0.5644671201814059, | |
| "eval_recall_micro": 0.6060606060606061, | |
| "eval_recall_weighted": 0.6060606060606061, | |
| "eval_runtime": 1.6837, | |
| "eval_samples_per_second": 78.398, | |
| "eval_steps_per_second": 10.097, | |
| "step": 1624 | |
| }, | |
| { | |
| "epoch": 28.03448275862069, | |
| "grad_norm": 0.1886104792356491, | |
| "learning_rate": 2.2112753147235906e-05, | |
| "loss": 0.0641, | |
| "step": 1626 | |
| }, | |
| { | |
| "epoch": 28.06896551724138, | |
| "grad_norm": 1.1970809698104858, | |
| "learning_rate": 2.2003284072249592e-05, | |
| "loss": 0.1138, | |
| "step": 1628 | |
| }, | |
| { | |
| "epoch": 28.103448275862068, | |
| "grad_norm": 0.11415582150220871, | |
| "learning_rate": 2.1893814997263272e-05, | |
| "loss": 0.0359, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 28.137931034482758, | |
| "grad_norm": 0.26719969511032104, | |
| "learning_rate": 2.1784345922276958e-05, | |
| "loss": 0.035, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 28.17241379310345, | |
| "grad_norm": 7.150755882263184, | |
| "learning_rate": 2.1674876847290644e-05, | |
| "loss": 0.0961, | |
| "step": 1634 | |
| }, | |
| { | |
| "epoch": 28.20689655172414, | |
| "grad_norm": 0.15368057787418365, | |
| "learning_rate": 2.1565407772304324e-05, | |
| "loss": 0.0322, | |
| "step": 1636 | |
| }, | |
| { | |
| "epoch": 28.24137931034483, | |
| "grad_norm": 1.2831923961639404, | |
| "learning_rate": 2.145593869731801e-05, | |
| "loss": 0.0906, | |
| "step": 1638 | |
| }, | |
| { | |
| "epoch": 28.275862068965516, | |
| "grad_norm": 0.44488441944122314, | |
| "learning_rate": 2.1346469622331693e-05, | |
| "loss": 0.0373, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 28.310344827586206, | |
| "grad_norm": 0.1512623429298401, | |
| "learning_rate": 2.1237000547345376e-05, | |
| "loss": 0.0398, | |
| "step": 1642 | |
| }, | |
| { | |
| "epoch": 28.344827586206897, | |
| "grad_norm": 15.000018119812012, | |
| "learning_rate": 2.1127531472359062e-05, | |
| "loss": 0.1686, | |
| "step": 1644 | |
| }, | |
| { | |
| "epoch": 28.379310344827587, | |
| "grad_norm": 0.198665052652359, | |
| "learning_rate": 2.101806239737274e-05, | |
| "loss": 0.0306, | |
| "step": 1646 | |
| }, | |
| { | |
| "epoch": 28.413793103448278, | |
| "grad_norm": 1.9664894342422485, | |
| "learning_rate": 2.0908593322386428e-05, | |
| "loss": 0.0556, | |
| "step": 1648 | |
| }, | |
| { | |
| "epoch": 28.448275862068964, | |
| "grad_norm": 0.1294247806072235, | |
| "learning_rate": 2.079912424740011e-05, | |
| "loss": 0.0341, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 28.482758620689655, | |
| "grad_norm": 0.22127756476402283, | |
| "learning_rate": 2.0689655172413793e-05, | |
| "loss": 0.0372, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 28.517241379310345, | |
| "grad_norm": 0.16517043113708496, | |
| "learning_rate": 2.058018609742748e-05, | |
| "loss": 0.0316, | |
| "step": 1654 | |
| }, | |
| { | |
| "epoch": 28.551724137931036, | |
| "grad_norm": 0.6368513107299805, | |
| "learning_rate": 2.047071702244116e-05, | |
| "loss": 0.0947, | |
| "step": 1656 | |
| }, | |
| { | |
| "epoch": 28.586206896551722, | |
| "grad_norm": 0.39542946219444275, | |
| "learning_rate": 2.0361247947454845e-05, | |
| "loss": 0.0356, | |
| "step": 1658 | |
| }, | |
| { | |
| "epoch": 28.620689655172413, | |
| "grad_norm": 0.3416825234889984, | |
| "learning_rate": 2.025177887246853e-05, | |
| "loss": 0.0625, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 28.655172413793103, | |
| "grad_norm": 1.6610913276672363, | |
| "learning_rate": 2.014230979748221e-05, | |
| "loss": 0.0392, | |
| "step": 1662 | |
| }, | |
| { | |
| "epoch": 28.689655172413794, | |
| "grad_norm": 0.4484691023826599, | |
| "learning_rate": 2.0032840722495897e-05, | |
| "loss": 0.0426, | |
| "step": 1664 | |
| }, | |
| { | |
| "epoch": 28.724137931034484, | |
| "grad_norm": 9.614035606384277, | |
| "learning_rate": 1.992337164750958e-05, | |
| "loss": 0.0708, | |
| "step": 1666 | |
| }, | |
| { | |
| "epoch": 28.75862068965517, | |
| "grad_norm": 0.17257235944271088, | |
| "learning_rate": 1.9813902572523263e-05, | |
| "loss": 0.0649, | |
| "step": 1668 | |
| }, | |
| { | |
| "epoch": 28.79310344827586, | |
| "grad_norm": 0.8080697059631348, | |
| "learning_rate": 1.970443349753695e-05, | |
| "loss": 0.0337, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 28.82758620689655, | |
| "grad_norm": 0.12455487251281738, | |
| "learning_rate": 1.959496442255063e-05, | |
| "loss": 0.0292, | |
| "step": 1672 | |
| }, | |
| { | |
| "epoch": 28.862068965517242, | |
| "grad_norm": 4.5023369789123535, | |
| "learning_rate": 1.9485495347564315e-05, | |
| "loss": 0.0603, | |
| "step": 1674 | |
| }, | |
| { | |
| "epoch": 28.896551724137932, | |
| "grad_norm": 7.025498390197754, | |
| "learning_rate": 1.9376026272577998e-05, | |
| "loss": 0.1706, | |
| "step": 1676 | |
| }, | |
| { | |
| "epoch": 28.93103448275862, | |
| "grad_norm": 5.604184627532959, | |
| "learning_rate": 1.926655719759168e-05, | |
| "loss": 0.3358, | |
| "step": 1678 | |
| }, | |
| { | |
| "epoch": 28.96551724137931, | |
| "grad_norm": 0.6545395255088806, | |
| "learning_rate": 1.9157088122605367e-05, | |
| "loss": 0.0469, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "grad_norm": 0.1356768012046814, | |
| "learning_rate": 1.9047619047619046e-05, | |
| "loss": 0.0269, | |
| "step": 1682 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5563089279356143, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6282606190642379, | |
| "eval_loss": 1.4752185344696045, | |
| "eval_precision_macro": 0.5686201153488518, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6350446686164523, | |
| "eval_recall_macro": 0.5514965986394558, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 1.7276, | |
| "eval_samples_per_second": 76.408, | |
| "eval_steps_per_second": 9.84, | |
| "step": 1682 | |
| }, | |
| { | |
| "epoch": 29.03448275862069, | |
| "grad_norm": 0.24686339497566223, | |
| "learning_rate": 1.8938149972632732e-05, | |
| "loss": 0.1751, | |
| "step": 1684 | |
| }, | |
| { | |
| "epoch": 29.06896551724138, | |
| "grad_norm": 14.179889678955078, | |
| "learning_rate": 1.8828680897646415e-05, | |
| "loss": 0.4523, | |
| "step": 1686 | |
| }, | |
| { | |
| "epoch": 29.103448275862068, | |
| "grad_norm": 0.6621524691581726, | |
| "learning_rate": 1.8719211822660098e-05, | |
| "loss": 0.0353, | |
| "step": 1688 | |
| }, | |
| { | |
| "epoch": 29.137931034482758, | |
| "grad_norm": 0.2671413719654083, | |
| "learning_rate": 1.8609742747673784e-05, | |
| "loss": 0.1784, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 29.17241379310345, | |
| "grad_norm": 0.11761974543333054, | |
| "learning_rate": 1.8500273672687464e-05, | |
| "loss": 0.0333, | |
| "step": 1692 | |
| }, | |
| { | |
| "epoch": 29.20689655172414, | |
| "grad_norm": 0.14111700654029846, | |
| "learning_rate": 1.839080459770115e-05, | |
| "loss": 0.0316, | |
| "step": 1694 | |
| }, | |
| { | |
| "epoch": 29.24137931034483, | |
| "grad_norm": 0.12016208469867706, | |
| "learning_rate": 1.8281335522714836e-05, | |
| "loss": 0.0484, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 29.275862068965516, | |
| "grad_norm": 0.16200877726078033, | |
| "learning_rate": 1.8171866447728516e-05, | |
| "loss": 0.0296, | |
| "step": 1698 | |
| }, | |
| { | |
| "epoch": 29.310344827586206, | |
| "grad_norm": 0.16708754003047943, | |
| "learning_rate": 1.8062397372742202e-05, | |
| "loss": 0.0322, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 29.344827586206897, | |
| "grad_norm": 0.1727229207754135, | |
| "learning_rate": 1.7952928297755885e-05, | |
| "loss": 0.0319, | |
| "step": 1702 | |
| }, | |
| { | |
| "epoch": 29.379310344827587, | |
| "grad_norm": 4.953850269317627, | |
| "learning_rate": 1.7843459222769568e-05, | |
| "loss": 0.1111, | |
| "step": 1704 | |
| }, | |
| { | |
| "epoch": 29.413793103448278, | |
| "grad_norm": 0.14302562177181244, | |
| "learning_rate": 1.7733990147783254e-05, | |
| "loss": 0.0494, | |
| "step": 1706 | |
| }, | |
| { | |
| "epoch": 29.448275862068964, | |
| "grad_norm": 0.39635199308395386, | |
| "learning_rate": 1.7624521072796933e-05, | |
| "loss": 0.0708, | |
| "step": 1708 | |
| }, | |
| { | |
| "epoch": 29.482758620689655, | |
| "grad_norm": 0.4933320879936218, | |
| "learning_rate": 1.751505199781062e-05, | |
| "loss": 0.0564, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 29.517241379310345, | |
| "grad_norm": 3.0435335636138916, | |
| "learning_rate": 1.7405582922824302e-05, | |
| "loss": 0.0757, | |
| "step": 1712 | |
| }, | |
| { | |
| "epoch": 29.551724137931036, | |
| "grad_norm": 0.11169429123401642, | |
| "learning_rate": 1.7296113847837985e-05, | |
| "loss": 0.0322, | |
| "step": 1714 | |
| }, | |
| { | |
| "epoch": 29.586206896551722, | |
| "grad_norm": 0.19250567257404327, | |
| "learning_rate": 1.718664477285167e-05, | |
| "loss": 0.0339, | |
| "step": 1716 | |
| }, | |
| { | |
| "epoch": 29.620689655172413, | |
| "grad_norm": 15.015722274780273, | |
| "learning_rate": 1.7077175697865354e-05, | |
| "loss": 0.1737, | |
| "step": 1718 | |
| }, | |
| { | |
| "epoch": 29.655172413793103, | |
| "grad_norm": 0.593941330909729, | |
| "learning_rate": 1.6967706622879037e-05, | |
| "loss": 0.0301, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 29.689655172413794, | |
| "grad_norm": 0.10976947844028473, | |
| "learning_rate": 1.6858237547892723e-05, | |
| "loss": 0.0442, | |
| "step": 1722 | |
| }, | |
| { | |
| "epoch": 29.724137931034484, | |
| "grad_norm": 29.424667358398438, | |
| "learning_rate": 1.6748768472906403e-05, | |
| "loss": 0.1292, | |
| "step": 1724 | |
| }, | |
| { | |
| "epoch": 29.75862068965517, | |
| "grad_norm": 0.1969483345746994, | |
| "learning_rate": 1.663929939792009e-05, | |
| "loss": 0.0394, | |
| "step": 1726 | |
| }, | |
| { | |
| "epoch": 29.79310344827586, | |
| "grad_norm": 0.39891931414604187, | |
| "learning_rate": 1.6529830322933772e-05, | |
| "loss": 0.0378, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 29.82758620689655, | |
| "grad_norm": 2.7617032527923584, | |
| "learning_rate": 1.6420361247947455e-05, | |
| "loss": 0.0351, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 29.862068965517242, | |
| "grad_norm": 4.983090400695801, | |
| "learning_rate": 1.631089217296114e-05, | |
| "loss": 0.0676, | |
| "step": 1732 | |
| }, | |
| { | |
| "epoch": 29.896551724137932, | |
| "grad_norm": 0.8957940340042114, | |
| "learning_rate": 1.620142309797482e-05, | |
| "loss": 0.0957, | |
| "step": 1734 | |
| }, | |
| { | |
| "epoch": 29.93103448275862, | |
| "grad_norm": 0.19039399921894073, | |
| "learning_rate": 1.6091954022988507e-05, | |
| "loss": 0.0263, | |
| "step": 1736 | |
| }, | |
| { | |
| "epoch": 29.96551724137931, | |
| "grad_norm": 0.091913141310215, | |
| "learning_rate": 1.598248494800219e-05, | |
| "loss": 0.0269, | |
| "step": 1738 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "grad_norm": 0.10125374048948288, | |
| "learning_rate": 1.5873015873015872e-05, | |
| "loss": 0.0267, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_accuracy": 0.6136363636363636, | |
| "eval_f1_macro": 0.562093981777284, | |
| "eval_f1_micro": 0.6136363636363636, | |
| "eval_f1_weighted": 0.6142366070214389, | |
| "eval_loss": 1.5352805852890015, | |
| "eval_precision_macro": 0.5859041645331968, | |
| "eval_precision_micro": 0.6136363636363636, | |
| "eval_precision_weighted": 0.6323715424857898, | |
| "eval_recall_macro": 0.5564550264550264, | |
| "eval_recall_micro": 0.6136363636363636, | |
| "eval_recall_weighted": 0.6136363636363636, | |
| "eval_runtime": 1.6762, | |
| "eval_samples_per_second": 78.751, | |
| "eval_steps_per_second": 10.142, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 30.03448275862069, | |
| "grad_norm": 0.1440192312002182, | |
| "learning_rate": 1.576354679802956e-05, | |
| "loss": 0.0348, | |
| "step": 1742 | |
| }, | |
| { | |
| "epoch": 30.06896551724138, | |
| "grad_norm": 0.8625943064689636, | |
| "learning_rate": 1.565407772304324e-05, | |
| "loss": 0.0343, | |
| "step": 1744 | |
| }, | |
| { | |
| "epoch": 30.103448275862068, | |
| "grad_norm": 0.2873701751232147, | |
| "learning_rate": 1.5544608648056924e-05, | |
| "loss": 0.0709, | |
| "step": 1746 | |
| }, | |
| { | |
| "epoch": 30.137931034482758, | |
| "grad_norm": 3.517399787902832, | |
| "learning_rate": 1.543513957307061e-05, | |
| "loss": 0.0444, | |
| "step": 1748 | |
| }, | |
| { | |
| "epoch": 30.17241379310345, | |
| "grad_norm": 0.18623045086860657, | |
| "learning_rate": 1.532567049808429e-05, | |
| "loss": 0.0292, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 30.20689655172414, | |
| "grad_norm": 0.24593006074428558, | |
| "learning_rate": 1.5216201423097976e-05, | |
| "loss": 0.0281, | |
| "step": 1752 | |
| }, | |
| { | |
| "epoch": 30.24137931034483, | |
| "grad_norm": 13.92945384979248, | |
| "learning_rate": 1.510673234811166e-05, | |
| "loss": 0.1024, | |
| "step": 1754 | |
| }, | |
| { | |
| "epoch": 30.275862068965516, | |
| "grad_norm": 0.12701399624347687, | |
| "learning_rate": 1.4997263273125342e-05, | |
| "loss": 0.0546, | |
| "step": 1756 | |
| }, | |
| { | |
| "epoch": 30.310344827586206, | |
| "grad_norm": 0.2253095507621765, | |
| "learning_rate": 1.4887794198139026e-05, | |
| "loss": 0.0313, | |
| "step": 1758 | |
| }, | |
| { | |
| "epoch": 30.344827586206897, | |
| "grad_norm": 0.18442504107952118, | |
| "learning_rate": 1.4778325123152711e-05, | |
| "loss": 0.045, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 30.379310344827587, | |
| "grad_norm": 0.8345377445220947, | |
| "learning_rate": 1.4668856048166394e-05, | |
| "loss": 0.0872, | |
| "step": 1762 | |
| }, | |
| { | |
| "epoch": 30.413793103448278, | |
| "grad_norm": 0.16499698162078857, | |
| "learning_rate": 1.4559386973180078e-05, | |
| "loss": 0.0521, | |
| "step": 1764 | |
| }, | |
| { | |
| "epoch": 30.448275862068964, | |
| "grad_norm": 29.77411460876465, | |
| "learning_rate": 1.444991789819376e-05, | |
| "loss": 0.075, | |
| "step": 1766 | |
| }, | |
| { | |
| "epoch": 30.482758620689655, | |
| "grad_norm": 13.939007759094238, | |
| "learning_rate": 1.4340448823207444e-05, | |
| "loss": 0.2595, | |
| "step": 1768 | |
| }, | |
| { | |
| "epoch": 30.517241379310345, | |
| "grad_norm": 0.110160693526268, | |
| "learning_rate": 1.4230979748221129e-05, | |
| "loss": 0.0255, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 30.551724137931036, | |
| "grad_norm": 0.17738622426986694, | |
| "learning_rate": 1.4121510673234811e-05, | |
| "loss": 0.0263, | |
| "step": 1772 | |
| }, | |
| { | |
| "epoch": 30.586206896551722, | |
| "grad_norm": 0.12670966982841492, | |
| "learning_rate": 1.4012041598248496e-05, | |
| "loss": 0.0309, | |
| "step": 1774 | |
| }, | |
| { | |
| "epoch": 30.620689655172413, | |
| "grad_norm": 0.08485020697116852, | |
| "learning_rate": 1.3902572523262177e-05, | |
| "loss": 0.0237, | |
| "step": 1776 | |
| }, | |
| { | |
| "epoch": 30.655172413793103, | |
| "grad_norm": 0.1434456706047058, | |
| "learning_rate": 1.3793103448275863e-05, | |
| "loss": 0.03, | |
| "step": 1778 | |
| }, | |
| { | |
| "epoch": 30.689655172413794, | |
| "grad_norm": 0.9384951591491699, | |
| "learning_rate": 1.3683634373289548e-05, | |
| "loss": 0.0309, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 30.724137931034484, | |
| "grad_norm": 1.0260897874832153, | |
| "learning_rate": 1.3574165298303229e-05, | |
| "loss": 0.0875, | |
| "step": 1782 | |
| }, | |
| { | |
| "epoch": 30.75862068965517, | |
| "grad_norm": 0.11444691568613052, | |
| "learning_rate": 1.3464696223316914e-05, | |
| "loss": 0.318, | |
| "step": 1784 | |
| }, | |
| { | |
| "epoch": 30.79310344827586, | |
| "grad_norm": 0.163078173995018, | |
| "learning_rate": 1.3355227148330598e-05, | |
| "loss": 0.0263, | |
| "step": 1786 | |
| }, | |
| { | |
| "epoch": 30.82758620689655, | |
| "grad_norm": 30.215320587158203, | |
| "learning_rate": 1.3245758073344281e-05, | |
| "loss": 0.1381, | |
| "step": 1788 | |
| }, | |
| { | |
| "epoch": 30.862068965517242, | |
| "grad_norm": 0.09894564747810364, | |
| "learning_rate": 1.3136288998357965e-05, | |
| "loss": 0.0427, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 30.896551724137932, | |
| "grad_norm": 0.15414094924926758, | |
| "learning_rate": 1.3026819923371647e-05, | |
| "loss": 0.0362, | |
| "step": 1792 | |
| }, | |
| { | |
| "epoch": 30.93103448275862, | |
| "grad_norm": 0.653509795665741, | |
| "learning_rate": 1.2917350848385331e-05, | |
| "loss": 0.0336, | |
| "step": 1794 | |
| }, | |
| { | |
| "epoch": 30.96551724137931, | |
| "grad_norm": 0.1620655208826065, | |
| "learning_rate": 1.2807881773399016e-05, | |
| "loss": 0.0314, | |
| "step": 1796 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "grad_norm": 0.39400652050971985, | |
| "learning_rate": 1.2698412698412699e-05, | |
| "loss": 0.1094, | |
| "step": 1798 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5912410632947628, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6529064358109834, | |
| "eval_loss": 1.5125658512115479, | |
| "eval_precision_macro": 0.602827742217164, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6604101388171674, | |
| "eval_recall_macro": 0.5866893424036281, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 1.7146, | |
| "eval_samples_per_second": 76.988, | |
| "eval_steps_per_second": 9.915, | |
| "step": 1798 | |
| }, | |
| { | |
| "epoch": 31.03448275862069, | |
| "grad_norm": 0.20587500929832458, | |
| "learning_rate": 1.2588943623426383e-05, | |
| "loss": 0.0298, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 31.06896551724138, | |
| "grad_norm": 0.27925363183021545, | |
| "learning_rate": 1.2479474548440066e-05, | |
| "loss": 0.0316, | |
| "step": 1802 | |
| }, | |
| { | |
| "epoch": 31.103448275862068, | |
| "grad_norm": 0.110267773270607, | |
| "learning_rate": 1.2370005473453749e-05, | |
| "loss": 0.0252, | |
| "step": 1804 | |
| }, | |
| { | |
| "epoch": 31.137931034482758, | |
| "grad_norm": 0.7436184883117676, | |
| "learning_rate": 1.2260536398467433e-05, | |
| "loss": 0.0288, | |
| "step": 1806 | |
| }, | |
| { | |
| "epoch": 31.17241379310345, | |
| "grad_norm": 0.1437726467847824, | |
| "learning_rate": 1.2151067323481118e-05, | |
| "loss": 0.0645, | |
| "step": 1808 | |
| }, | |
| { | |
| "epoch": 31.20689655172414, | |
| "grad_norm": 0.10505229979753494, | |
| "learning_rate": 1.20415982484948e-05, | |
| "loss": 0.7232, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 31.24137931034483, | |
| "grad_norm": 0.2323630303144455, | |
| "learning_rate": 1.1932129173508484e-05, | |
| "loss": 0.031, | |
| "step": 1812 | |
| }, | |
| { | |
| "epoch": 31.275862068965516, | |
| "grad_norm": 0.15194958448410034, | |
| "learning_rate": 1.1822660098522168e-05, | |
| "loss": 0.0301, | |
| "step": 1814 | |
| }, | |
| { | |
| "epoch": 31.310344827586206, | |
| "grad_norm": 0.29873228073120117, | |
| "learning_rate": 1.1713191023535853e-05, | |
| "loss": 0.0289, | |
| "step": 1816 | |
| }, | |
| { | |
| "epoch": 31.344827586206897, | |
| "grad_norm": 0.08876697719097137, | |
| "learning_rate": 1.1603721948549535e-05, | |
| "loss": 0.0267, | |
| "step": 1818 | |
| }, | |
| { | |
| "epoch": 31.379310344827587, | |
| "grad_norm": 0.5476172566413879, | |
| "learning_rate": 1.1494252873563218e-05, | |
| "loss": 0.0398, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 31.413793103448278, | |
| "grad_norm": 6.674413204193115, | |
| "learning_rate": 1.1384783798576903e-05, | |
| "loss": 0.0472, | |
| "step": 1822 | |
| }, | |
| { | |
| "epoch": 31.448275862068964, | |
| "grad_norm": 0.09850436449050903, | |
| "learning_rate": 1.1275314723590587e-05, | |
| "loss": 0.0296, | |
| "step": 1824 | |
| }, | |
| { | |
| "epoch": 31.482758620689655, | |
| "grad_norm": 0.5808874368667603, | |
| "learning_rate": 1.116584564860427e-05, | |
| "loss": 0.0507, | |
| "step": 1826 | |
| }, | |
| { | |
| "epoch": 31.517241379310345, | |
| "grad_norm": 0.10986245423555374, | |
| "learning_rate": 1.1056376573617953e-05, | |
| "loss": 0.0271, | |
| "step": 1828 | |
| }, | |
| { | |
| "epoch": 31.551724137931036, | |
| "grad_norm": 0.17757023870944977, | |
| "learning_rate": 1.0946907498631636e-05, | |
| "loss": 0.0306, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 31.586206896551722, | |
| "grad_norm": 0.1803562492132187, | |
| "learning_rate": 1.0837438423645322e-05, | |
| "loss": 0.0263, | |
| "step": 1832 | |
| }, | |
| { | |
| "epoch": 31.620689655172413, | |
| "grad_norm": 0.5573262572288513, | |
| "learning_rate": 1.0727969348659005e-05, | |
| "loss": 0.0277, | |
| "step": 1834 | |
| }, | |
| { | |
| "epoch": 31.655172413793103, | |
| "grad_norm": 0.9371552467346191, | |
| "learning_rate": 1.0618500273672688e-05, | |
| "loss": 0.0396, | |
| "step": 1836 | |
| }, | |
| { | |
| "epoch": 31.689655172413794, | |
| "grad_norm": 0.5467616319656372, | |
| "learning_rate": 1.050903119868637e-05, | |
| "loss": 0.035, | |
| "step": 1838 | |
| }, | |
| { | |
| "epoch": 31.724137931034484, | |
| "grad_norm": 0.18951688706874847, | |
| "learning_rate": 1.0399562123700055e-05, | |
| "loss": 0.0292, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 31.75862068965517, | |
| "grad_norm": 0.13354890048503876, | |
| "learning_rate": 1.029009304871374e-05, | |
| "loss": 0.0244, | |
| "step": 1842 | |
| }, | |
| { | |
| "epoch": 31.79310344827586, | |
| "grad_norm": 0.436275452375412, | |
| "learning_rate": 1.0180623973727423e-05, | |
| "loss": 0.0301, | |
| "step": 1844 | |
| }, | |
| { | |
| "epoch": 31.82758620689655, | |
| "grad_norm": 0.963607907295227, | |
| "learning_rate": 1.0071154898741105e-05, | |
| "loss": 0.0286, | |
| "step": 1846 | |
| }, | |
| { | |
| "epoch": 31.862068965517242, | |
| "grad_norm": 5.803387641906738, | |
| "learning_rate": 9.96168582375479e-06, | |
| "loss": 0.0649, | |
| "step": 1848 | |
| }, | |
| { | |
| "epoch": 31.896551724137932, | |
| "grad_norm": 0.3790993392467499, | |
| "learning_rate": 9.852216748768475e-06, | |
| "loss": 0.0327, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 31.93103448275862, | |
| "grad_norm": 0.08771317452192307, | |
| "learning_rate": 9.742747673782157e-06, | |
| "loss": 0.0247, | |
| "step": 1852 | |
| }, | |
| { | |
| "epoch": 31.96551724137931, | |
| "grad_norm": 0.15166075527668, | |
| "learning_rate": 9.63327859879584e-06, | |
| "loss": 0.0284, | |
| "step": 1854 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "grad_norm": 0.09096769243478775, | |
| "learning_rate": 9.523809523809523e-06, | |
| "loss": 0.0243, | |
| "step": 1856 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_accuracy": 0.6590909090909091, | |
| "eval_f1_macro": 0.5985429204941399, | |
| "eval_f1_micro": 0.6590909090909091, | |
| "eval_f1_weighted": 0.6563486864207485, | |
| "eval_loss": 1.490037202835083, | |
| "eval_precision_macro": 0.6103033255742615, | |
| "eval_precision_micro": 0.6590909090909091, | |
| "eval_precision_weighted": 0.6603901845281156, | |
| "eval_recall_macro": 0.5934920634920635, | |
| "eval_recall_micro": 0.6590909090909091, | |
| "eval_recall_weighted": 0.6590909090909091, | |
| "eval_runtime": 1.727, | |
| "eval_samples_per_second": 76.433, | |
| "eval_steps_per_second": 9.844, | |
| "step": 1856 | |
| }, | |
| { | |
| "epoch": 32.03448275862069, | |
| "grad_norm": 0.18323098123073578, | |
| "learning_rate": 9.414340448823208e-06, | |
| "loss": 0.0304, | |
| "step": 1858 | |
| }, | |
| { | |
| "epoch": 32.06896551724138, | |
| "grad_norm": 0.15102934837341309, | |
| "learning_rate": 9.304871373836892e-06, | |
| "loss": 0.0247, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 32.10344827586207, | |
| "grad_norm": 0.151427760720253, | |
| "learning_rate": 9.195402298850575e-06, | |
| "loss": 0.0246, | |
| "step": 1862 | |
| }, | |
| { | |
| "epoch": 32.13793103448276, | |
| "grad_norm": 0.10605157166719437, | |
| "learning_rate": 9.085933223864258e-06, | |
| "loss": 0.0282, | |
| "step": 1864 | |
| }, | |
| { | |
| "epoch": 32.172413793103445, | |
| "grad_norm": 0.0751444399356842, | |
| "learning_rate": 8.976464148877942e-06, | |
| "loss": 0.023, | |
| "step": 1866 | |
| }, | |
| { | |
| "epoch": 32.206896551724135, | |
| "grad_norm": 0.13312040269374847, | |
| "learning_rate": 8.866995073891627e-06, | |
| "loss": 0.0256, | |
| "step": 1868 | |
| }, | |
| { | |
| "epoch": 32.241379310344826, | |
| "grad_norm": 0.10063087195158005, | |
| "learning_rate": 8.75752599890531e-06, | |
| "loss": 0.022, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 32.275862068965516, | |
| "grad_norm": 0.11364129185676575, | |
| "learning_rate": 8.648056923918993e-06, | |
| "loss": 0.0968, | |
| "step": 1872 | |
| }, | |
| { | |
| "epoch": 32.310344827586206, | |
| "grad_norm": 0.09259811043739319, | |
| "learning_rate": 8.538587848932677e-06, | |
| "loss": 0.0244, | |
| "step": 1874 | |
| }, | |
| { | |
| "epoch": 32.3448275862069, | |
| "grad_norm": 3.0221097469329834, | |
| "learning_rate": 8.429118773946362e-06, | |
| "loss": 0.0491, | |
| "step": 1876 | |
| }, | |
| { | |
| "epoch": 32.37931034482759, | |
| "grad_norm": 0.09685959666967392, | |
| "learning_rate": 8.319649698960045e-06, | |
| "loss": 0.0244, | |
| "step": 1878 | |
| }, | |
| { | |
| "epoch": 32.41379310344828, | |
| "grad_norm": 0.13511282205581665, | |
| "learning_rate": 8.210180623973727e-06, | |
| "loss": 0.0271, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 32.44827586206897, | |
| "grad_norm": 0.12241239845752716, | |
| "learning_rate": 8.10071154898741e-06, | |
| "loss": 0.0272, | |
| "step": 1882 | |
| }, | |
| { | |
| "epoch": 32.48275862068966, | |
| "grad_norm": 0.1032300814986229, | |
| "learning_rate": 7.991242474001095e-06, | |
| "loss": 0.0229, | |
| "step": 1884 | |
| }, | |
| { | |
| "epoch": 32.51724137931034, | |
| "grad_norm": 0.4652920961380005, | |
| "learning_rate": 7.88177339901478e-06, | |
| "loss": 0.0273, | |
| "step": 1886 | |
| }, | |
| { | |
| "epoch": 32.55172413793103, | |
| "grad_norm": 0.28032049536705017, | |
| "learning_rate": 7.772304324028462e-06, | |
| "loss": 0.0298, | |
| "step": 1888 | |
| }, | |
| { | |
| "epoch": 32.58620689655172, | |
| "grad_norm": 0.6975933313369751, | |
| "learning_rate": 7.662835249042145e-06, | |
| "loss": 0.0293, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 32.62068965517241, | |
| "grad_norm": 0.1151973232626915, | |
| "learning_rate": 7.55336617405583e-06, | |
| "loss": 0.0233, | |
| "step": 1892 | |
| }, | |
| { | |
| "epoch": 32.6551724137931, | |
| "grad_norm": 16.620948791503906, | |
| "learning_rate": 7.443897099069513e-06, | |
| "loss": 0.1329, | |
| "step": 1894 | |
| }, | |
| { | |
| "epoch": 32.689655172413794, | |
| "grad_norm": 0.11886937916278839, | |
| "learning_rate": 7.334428024083197e-06, | |
| "loss": 0.0215, | |
| "step": 1896 | |
| }, | |
| { | |
| "epoch": 32.724137931034484, | |
| "grad_norm": 0.699009120464325, | |
| "learning_rate": 7.22495894909688e-06, | |
| "loss": 0.0262, | |
| "step": 1898 | |
| }, | |
| { | |
| "epoch": 32.758620689655174, | |
| "grad_norm": 0.5381059050559998, | |
| "learning_rate": 7.115489874110564e-06, | |
| "loss": 0.0339, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 32.793103448275865, | |
| "grad_norm": 0.09165710210800171, | |
| "learning_rate": 7.006020799124248e-06, | |
| "loss": 0.0237, | |
| "step": 1902 | |
| }, | |
| { | |
| "epoch": 32.827586206896555, | |
| "grad_norm": 0.09409786015748978, | |
| "learning_rate": 6.896551724137932e-06, | |
| "loss": 0.0376, | |
| "step": 1904 | |
| }, | |
| { | |
| "epoch": 32.86206896551724, | |
| "grad_norm": 0.11274611204862595, | |
| "learning_rate": 6.7870826491516145e-06, | |
| "loss": 0.0268, | |
| "step": 1906 | |
| }, | |
| { | |
| "epoch": 32.89655172413793, | |
| "grad_norm": 0.7047907710075378, | |
| "learning_rate": 6.677613574165299e-06, | |
| "loss": 0.0679, | |
| "step": 1908 | |
| }, | |
| { | |
| "epoch": 32.93103448275862, | |
| "grad_norm": 0.11477731913328171, | |
| "learning_rate": 6.568144499178983e-06, | |
| "loss": 0.0208, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 32.96551724137931, | |
| "grad_norm": 0.09211686998605728, | |
| "learning_rate": 6.458675424192666e-06, | |
| "loss": 0.2335, | |
| "step": 1912 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "grad_norm": 0.13210824131965637, | |
| "learning_rate": 6.349206349206349e-06, | |
| "loss": 0.0366, | |
| "step": 1914 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_accuracy": 0.6893939393939394, | |
| "eval_f1_macro": 0.6274757091289592, | |
| "eval_f1_micro": 0.6893939393939394, | |
| "eval_f1_weighted": 0.6850690417795446, | |
| "eval_loss": 1.467975378036499, | |
| "eval_precision_macro": 0.6369170238628366, | |
| "eval_precision_micro": 0.6893939393939394, | |
| "eval_precision_weighted": 0.6855083526713621, | |
| "eval_recall_macro": 0.6241043083900227, | |
| "eval_recall_micro": 0.6893939393939394, | |
| "eval_recall_weighted": 0.6893939393939394, | |
| "eval_runtime": 1.744, | |
| "eval_samples_per_second": 75.687, | |
| "eval_steps_per_second": 9.748, | |
| "step": 1914 | |
| }, | |
| { | |
| "epoch": 33.03448275862069, | |
| "grad_norm": 0.13359691202640533, | |
| "learning_rate": 6.239737274220033e-06, | |
| "loss": 0.0275, | |
| "step": 1916 | |
| }, | |
| { | |
| "epoch": 33.06896551724138, | |
| "grad_norm": 0.2554221749305725, | |
| "learning_rate": 6.130268199233717e-06, | |
| "loss": 0.0298, | |
| "step": 1918 | |
| }, | |
| { | |
| "epoch": 33.10344827586207, | |
| "grad_norm": 0.1016729325056076, | |
| "learning_rate": 6.0207991242474e-06, | |
| "loss": 0.0271, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 33.13793103448276, | |
| "grad_norm": 0.0994502529501915, | |
| "learning_rate": 5.911330049261084e-06, | |
| "loss": 0.0238, | |
| "step": 1922 | |
| }, | |
| { | |
| "epoch": 33.172413793103445, | |
| "grad_norm": 0.08981093764305115, | |
| "learning_rate": 5.801860974274768e-06, | |
| "loss": 0.0206, | |
| "step": 1924 | |
| }, | |
| { | |
| "epoch": 33.206896551724135, | |
| "grad_norm": 0.7163501977920532, | |
| "learning_rate": 5.692391899288451e-06, | |
| "loss": 0.0307, | |
| "step": 1926 | |
| }, | |
| { | |
| "epoch": 33.241379310344826, | |
| "grad_norm": 2.008106231689453, | |
| "learning_rate": 5.582922824302135e-06, | |
| "loss": 0.0682, | |
| "step": 1928 | |
| }, | |
| { | |
| "epoch": 33.275862068965516, | |
| "grad_norm": 0.15033219754695892, | |
| "learning_rate": 5.473453749315818e-06, | |
| "loss": 0.0248, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 33.310344827586206, | |
| "grad_norm": 0.5542562007904053, | |
| "learning_rate": 5.3639846743295025e-06, | |
| "loss": 0.0236, | |
| "step": 1932 | |
| }, | |
| { | |
| "epoch": 33.3448275862069, | |
| "grad_norm": 0.19591717422008514, | |
| "learning_rate": 5.254515599343185e-06, | |
| "loss": 0.0299, | |
| "step": 1934 | |
| }, | |
| { | |
| "epoch": 33.37931034482759, | |
| "grad_norm": 0.12007873505353928, | |
| "learning_rate": 5.14504652435687e-06, | |
| "loss": 0.0249, | |
| "step": 1936 | |
| }, | |
| { | |
| "epoch": 33.41379310344828, | |
| "grad_norm": 0.6954270005226135, | |
| "learning_rate": 5.035577449370553e-06, | |
| "loss": 0.0287, | |
| "step": 1938 | |
| }, | |
| { | |
| "epoch": 33.44827586206897, | |
| "grad_norm": 1.665319800376892, | |
| "learning_rate": 4.926108374384237e-06, | |
| "loss": 0.0373, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 33.48275862068966, | |
| "grad_norm": 0.8834956884384155, | |
| "learning_rate": 4.81663929939792e-06, | |
| "loss": 0.1, | |
| "step": 1942 | |
| }, | |
| { | |
| "epoch": 33.51724137931034, | |
| "grad_norm": 0.09338194131851196, | |
| "learning_rate": 4.707170224411604e-06, | |
| "loss": 0.0281, | |
| "step": 1944 | |
| }, | |
| { | |
| "epoch": 33.55172413793103, | |
| "grad_norm": 0.11382108926773071, | |
| "learning_rate": 4.5977011494252875e-06, | |
| "loss": 0.0255, | |
| "step": 1946 | |
| }, | |
| { | |
| "epoch": 33.58620689655172, | |
| "grad_norm": 0.17863136529922485, | |
| "learning_rate": 4.488232074438971e-06, | |
| "loss": 0.0271, | |
| "step": 1948 | |
| }, | |
| { | |
| "epoch": 33.62068965517241, | |
| "grad_norm": 0.7330889105796814, | |
| "learning_rate": 4.378762999452655e-06, | |
| "loss": 0.0244, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 33.6551724137931, | |
| "grad_norm": 0.08643897622823715, | |
| "learning_rate": 4.2692939244663386e-06, | |
| "loss": 0.0217, | |
| "step": 1952 | |
| }, | |
| { | |
| "epoch": 33.689655172413794, | |
| "grad_norm": 0.09883402287960052, | |
| "learning_rate": 4.159824849480022e-06, | |
| "loss": 0.0238, | |
| "step": 1954 | |
| }, | |
| { | |
| "epoch": 33.724137931034484, | |
| "grad_norm": 0.11817536503076553, | |
| "learning_rate": 4.050355774493705e-06, | |
| "loss": 0.0254, | |
| "step": 1956 | |
| }, | |
| { | |
| "epoch": 33.758620689655174, | |
| "grad_norm": 0.09555190801620483, | |
| "learning_rate": 3.94088669950739e-06, | |
| "loss": 0.0211, | |
| "step": 1958 | |
| }, | |
| { | |
| "epoch": 33.793103448275865, | |
| "grad_norm": 18.8179931640625, | |
| "learning_rate": 3.8314176245210725e-06, | |
| "loss": 0.1301, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 33.827586206896555, | |
| "grad_norm": 0.10722629725933075, | |
| "learning_rate": 3.7219485495347566e-06, | |
| "loss": 0.0984, | |
| "step": 1962 | |
| }, | |
| { | |
| "epoch": 33.86206896551724, | |
| "grad_norm": 7.371537208557129, | |
| "learning_rate": 3.61247947454844e-06, | |
| "loss": 0.0517, | |
| "step": 1964 | |
| }, | |
| { | |
| "epoch": 33.89655172413793, | |
| "grad_norm": 0.13865648210048676, | |
| "learning_rate": 3.503010399562124e-06, | |
| "loss": 0.0268, | |
| "step": 1966 | |
| }, | |
| { | |
| "epoch": 33.93103448275862, | |
| "grad_norm": 0.08597588539123535, | |
| "learning_rate": 3.3935413245758073e-06, | |
| "loss": 0.0238, | |
| "step": 1968 | |
| }, | |
| { | |
| "epoch": 33.96551724137931, | |
| "grad_norm": 0.10678116977214813, | |
| "learning_rate": 3.2840722495894914e-06, | |
| "loss": 0.0234, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "grad_norm": 0.17705120146274567, | |
| "learning_rate": 3.1746031746031746e-06, | |
| "loss": 0.0235, | |
| "step": 1972 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_accuracy": 0.6818181818181818, | |
| "eval_f1_macro": 0.6216473911055191, | |
| "eval_f1_micro": 0.6818181818181818, | |
| "eval_f1_weighted": 0.6795081669559204, | |
| "eval_loss": 1.477163553237915, | |
| "eval_precision_macro": 0.6323631514971935, | |
| "eval_precision_micro": 0.6818181818181818, | |
| "eval_precision_weighted": 0.6836445631033983, | |
| "eval_recall_macro": 0.6173015873015872, | |
| "eval_recall_micro": 0.6818181818181818, | |
| "eval_recall_weighted": 0.6818181818181818, | |
| "eval_runtime": 1.697, | |
| "eval_samples_per_second": 77.783, | |
| "eval_steps_per_second": 10.018, | |
| "step": 1972 | |
| }, | |
| { | |
| "epoch": 34.03448275862069, | |
| "grad_norm": 0.09892508387565613, | |
| "learning_rate": 3.0651340996168583e-06, | |
| "loss": 0.0225, | |
| "step": 1974 | |
| }, | |
| { | |
| "epoch": 34.06896551724138, | |
| "grad_norm": 0.09665203094482422, | |
| "learning_rate": 2.955665024630542e-06, | |
| "loss": 0.0312, | |
| "step": 1976 | |
| }, | |
| { | |
| "epoch": 34.10344827586207, | |
| "grad_norm": 0.09741330146789551, | |
| "learning_rate": 2.8461959496442257e-06, | |
| "loss": 0.0222, | |
| "step": 1978 | |
| }, | |
| { | |
| "epoch": 34.13793103448276, | |
| "grad_norm": 0.08742016553878784, | |
| "learning_rate": 2.736726874657909e-06, | |
| "loss": 0.1584, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 34.172413793103445, | |
| "grad_norm": 0.5281866192817688, | |
| "learning_rate": 2.6272577996715927e-06, | |
| "loss": 0.027, | |
| "step": 1982 | |
| }, | |
| { | |
| "epoch": 34.206896551724135, | |
| "grad_norm": 0.12132374942302704, | |
| "learning_rate": 2.5177887246852764e-06, | |
| "loss": 0.0298, | |
| "step": 1984 | |
| }, | |
| { | |
| "epoch": 34.241379310344826, | |
| "grad_norm": 0.0994863212108612, | |
| "learning_rate": 2.40831964969896e-06, | |
| "loss": 0.0251, | |
| "step": 1986 | |
| }, | |
| { | |
| "epoch": 34.275862068965516, | |
| "grad_norm": 11.063030242919922, | |
| "learning_rate": 2.2988505747126437e-06, | |
| "loss": 0.0678, | |
| "step": 1988 | |
| }, | |
| { | |
| "epoch": 34.310344827586206, | |
| "grad_norm": 0.30384892225265503, | |
| "learning_rate": 2.1893814997263274e-06, | |
| "loss": 0.0243, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 34.3448275862069, | |
| "grad_norm": 0.09084097295999527, | |
| "learning_rate": 2.079912424740011e-06, | |
| "loss": 0.0826, | |
| "step": 1992 | |
| }, | |
| { | |
| "epoch": 34.37931034482759, | |
| "grad_norm": 0.10241986811161041, | |
| "learning_rate": 1.970443349753695e-06, | |
| "loss": 0.0222, | |
| "step": 1994 | |
| }, | |
| { | |
| "epoch": 34.41379310344828, | |
| "grad_norm": 0.0835302397608757, | |
| "learning_rate": 1.8609742747673783e-06, | |
| "loss": 0.026, | |
| "step": 1996 | |
| }, | |
| { | |
| "epoch": 34.44827586206897, | |
| "grad_norm": 0.0980919674038887, | |
| "learning_rate": 1.751505199781062e-06, | |
| "loss": 0.0267, | |
| "step": 1998 | |
| }, | |
| { | |
| "epoch": 34.48275862068966, | |
| "grad_norm": 0.15463948249816895, | |
| "learning_rate": 1.6420361247947457e-06, | |
| "loss": 0.0259, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 34.51724137931034, | |
| "grad_norm": 5.89515495300293, | |
| "learning_rate": 1.5325670498084292e-06, | |
| "loss": 0.0348, | |
| "step": 2002 | |
| }, | |
| { | |
| "epoch": 34.55172413793103, | |
| "grad_norm": 0.08560877293348312, | |
| "learning_rate": 1.4230979748221129e-06, | |
| "loss": 0.0227, | |
| "step": 2004 | |
| }, | |
| { | |
| "epoch": 34.58620689655172, | |
| "grad_norm": 0.152249276638031, | |
| "learning_rate": 1.3136288998357963e-06, | |
| "loss": 0.0232, | |
| "step": 2006 | |
| }, | |
| { | |
| "epoch": 34.62068965517241, | |
| "grad_norm": 0.13286259770393372, | |
| "learning_rate": 1.20415982484948e-06, | |
| "loss": 0.0227, | |
| "step": 2008 | |
| }, | |
| { | |
| "epoch": 34.6551724137931, | |
| "grad_norm": 0.13158464431762695, | |
| "learning_rate": 1.0946907498631637e-06, | |
| "loss": 0.0221, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 34.689655172413794, | |
| "grad_norm": 0.07431570440530777, | |
| "learning_rate": 9.852216748768474e-07, | |
| "loss": 0.0268, | |
| "step": 2012 | |
| }, | |
| { | |
| "epoch": 34.724137931034484, | |
| "grad_norm": 0.11424873024225235, | |
| "learning_rate": 8.75752599890531e-07, | |
| "loss": 0.0236, | |
| "step": 2014 | |
| }, | |
| { | |
| "epoch": 34.758620689655174, | |
| "grad_norm": 0.5630809664726257, | |
| "learning_rate": 7.662835249042146e-07, | |
| "loss": 0.0301, | |
| "step": 2016 | |
| }, | |
| { | |
| "epoch": 34.793103448275865, | |
| "grad_norm": 0.15935018658638, | |
| "learning_rate": 6.568144499178982e-07, | |
| "loss": 0.0274, | |
| "step": 2018 | |
| }, | |
| { | |
| "epoch": 34.827586206896555, | |
| "grad_norm": 0.08231520652770996, | |
| "learning_rate": 5.473453749315819e-07, | |
| "loss": 0.022, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 34.86206896551724, | |
| "grad_norm": 0.28459644317626953, | |
| "learning_rate": 4.378762999452655e-07, | |
| "loss": 0.0256, | |
| "step": 2022 | |
| }, | |
| { | |
| "epoch": 34.89655172413793, | |
| "grad_norm": 0.11006944626569748, | |
| "learning_rate": 3.284072249589491e-07, | |
| "loss": 0.0702, | |
| "step": 2024 | |
| }, | |
| { | |
| "epoch": 34.93103448275862, | |
| "grad_norm": 1.2310376167297363, | |
| "learning_rate": 2.1893814997263275e-07, | |
| "loss": 0.025, | |
| "step": 2026 | |
| }, | |
| { | |
| "epoch": 34.96551724137931, | |
| "grad_norm": 0.082310751080513, | |
| "learning_rate": 1.0946907498631637e-07, | |
| "loss": 0.0244, | |
| "step": 2028 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "grad_norm": 0.20153437554836273, | |
| "learning_rate": 0.0, | |
| "loss": 0.0345, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_accuracy": 0.696969696969697, | |
| "eval_f1_macro": 0.6556072550024405, | |
| "eval_f1_micro": 0.696969696969697, | |
| "eval_f1_weighted": 0.6961160958984463, | |
| "eval_loss": 1.4754124879837036, | |
| "eval_precision_macro": 0.6722334315930374, | |
| "eval_precision_micro": 0.696969696969697, | |
| "eval_precision_weighted": 0.703819782620723, | |
| "eval_recall_macro": 0.6479138321995465, | |
| "eval_recall_micro": 0.696969696969697, | |
| "eval_recall_weighted": 0.696969696969697, | |
| "eval_runtime": 1.8533, | |
| "eval_samples_per_second": 71.223, | |
| "eval_steps_per_second": 9.173, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "step": 2030, | |
| "total_flos": 1.2531016253190758e+18, | |
| "train_loss": 0.7230312045223167, | |
| "train_runtime": 345.4127, | |
| "train_samples_per_second": 46.814, | |
| "train_steps_per_second": 5.877 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 2030, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 35, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.2531016253190758e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |