{ "best_global_step": 400, "best_metric": 0.9147540983606557, "best_model_checkpoint": "./modelka_bebro/checkpoint-400", "epoch": 20.0, "eval_steps": 20, "global_step": 860, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.023255813953488372, "grad_norm": 0.4809555113315582, "learning_rate": 0.0, "loss": 0.6913, "step": 1 }, { "epoch": 0.046511627906976744, "grad_norm": 0.4111059904098511, "learning_rate": 5.813953488372093e-07, "loss": 0.6908, "step": 2 }, { "epoch": 0.06976744186046512, "grad_norm": 0.4677865207195282, "learning_rate": 1.1627906976744186e-06, "loss": 0.6945, "step": 3 }, { "epoch": 0.09302325581395349, "grad_norm": 0.42459365725517273, "learning_rate": 1.744186046511628e-06, "loss": 0.6974, "step": 4 }, { "epoch": 0.11627906976744186, "grad_norm": 0.5455716252326965, "learning_rate": 2.325581395348837e-06, "loss": 0.6876, "step": 5 }, { "epoch": 0.13953488372093023, "grad_norm": 0.3835294842720032, "learning_rate": 2.9069767441860468e-06, "loss": 0.6882, "step": 6 }, { "epoch": 0.16279069767441862, "grad_norm": 0.6538524627685547, "learning_rate": 3.488372093023256e-06, "loss": 0.6888, "step": 7 }, { "epoch": 0.18604651162790697, "grad_norm": 0.40157362818717957, "learning_rate": 4.0697674418604655e-06, "loss": 0.6855, "step": 8 }, { "epoch": 0.20930232558139536, "grad_norm": 0.43001478910446167, "learning_rate": 4.651162790697674e-06, "loss": 0.6918, "step": 9 }, { "epoch": 0.23255813953488372, "grad_norm": 0.5613086819648743, "learning_rate": 5.232558139534884e-06, "loss": 0.6977, "step": 10 }, { "epoch": 0.2558139534883721, "grad_norm": 0.3894985020160675, "learning_rate": 5.8139534883720935e-06, "loss": 0.6847, "step": 11 }, { "epoch": 0.27906976744186046, "grad_norm": 0.4243035912513733, "learning_rate": 6.395348837209303e-06, "loss": 0.6934, "step": 12 }, { "epoch": 0.3023255813953488, "grad_norm": 0.5996053218841553, "learning_rate": 6.976744186046512e-06, "loss": 0.6966, "step": 13 }, { "epoch": 0.32558139534883723, "grad_norm": 0.5445254445075989, "learning_rate": 7.558139534883721e-06, "loss": 0.695, "step": 14 }, { "epoch": 0.3488372093023256, "grad_norm": 0.433915913105011, "learning_rate": 8.139534883720931e-06, "loss": 0.6855, "step": 15 }, { "epoch": 0.37209302325581395, "grad_norm": 0.43263381719589233, "learning_rate": 8.72093023255814e-06, "loss": 0.6901, "step": 16 }, { "epoch": 0.3953488372093023, "grad_norm": 0.5314894318580627, "learning_rate": 9.302325581395349e-06, "loss": 0.6966, "step": 17 }, { "epoch": 0.4186046511627907, "grad_norm": 0.3562123477458954, "learning_rate": 9.883720930232558e-06, "loss": 0.6877, "step": 18 }, { "epoch": 0.4418604651162791, "grad_norm": 0.4751898944377899, "learning_rate": 1.0465116279069768e-05, "loss": 0.6834, "step": 19 }, { "epoch": 0.46511627906976744, "grad_norm": 0.4273492395877838, "learning_rate": 1.1046511627906977e-05, "loss": 0.6885, "step": 20 }, { "epoch": 0.46511627906976744, "eval_accuracy": 0.5901639344262295, "eval_f1": 0.5314093976229397, "eval_loss": 0.68656325340271, "eval_precision": 0.6364057433997221, "eval_recall": 0.5761178599121219, "eval_runtime": 0.0751, "eval_samples_per_second": 4062.686, "eval_steps_per_second": 66.601, "step": 20 }, { "epoch": 0.4883720930232558, "grad_norm": 0.5205795764923096, "learning_rate": 1.1627906976744187e-05, "loss": 0.6908, "step": 21 }, { "epoch": 0.5116279069767442, "grad_norm": 0.4466778635978699, "learning_rate": 1.2209302325581395e-05, "loss": 0.6861, "step": 22 }, { "epoch": 0.5348837209302325, "grad_norm": 0.44298067688941956, "learning_rate": 1.2790697674418606e-05, "loss": 0.6832, "step": 23 }, { "epoch": 0.5581395348837209, "grad_norm": 0.39592233300209045, "learning_rate": 1.3372093023255814e-05, "loss": 0.6877, "step": 24 }, { "epoch": 0.5813953488372093, "grad_norm": 0.4810555875301361, "learning_rate": 1.3953488372093024e-05, "loss": 0.68, "step": 25 }, { "epoch": 0.6046511627906976, "grad_norm": 0.47530311346054077, "learning_rate": 1.4534883720930233e-05, "loss": 0.6837, "step": 26 }, { "epoch": 0.627906976744186, "grad_norm": 0.5110817551612854, "learning_rate": 1.5116279069767441e-05, "loss": 0.6908, "step": 27 }, { "epoch": 0.6511627906976745, "grad_norm": 0.44798141717910767, "learning_rate": 1.569767441860465e-05, "loss": 0.6756, "step": 28 }, { "epoch": 0.6744186046511628, "grad_norm": 0.4494743347167969, "learning_rate": 1.6279069767441862e-05, "loss": 0.6813, "step": 29 }, { "epoch": 0.6976744186046512, "grad_norm": 0.4449148178100586, "learning_rate": 1.686046511627907e-05, "loss": 0.6786, "step": 30 }, { "epoch": 0.7209302325581395, "grad_norm": 0.49257639050483704, "learning_rate": 1.744186046511628e-05, "loss": 0.6821, "step": 31 }, { "epoch": 0.7441860465116279, "grad_norm": 0.49307772517204285, "learning_rate": 1.802325581395349e-05, "loss": 0.6856, "step": 32 }, { "epoch": 0.7674418604651163, "grad_norm": 0.4002581536769867, "learning_rate": 1.8604651162790697e-05, "loss": 0.6804, "step": 33 }, { "epoch": 0.7906976744186046, "grad_norm": 0.44210293889045715, "learning_rate": 1.918604651162791e-05, "loss": 0.6739, "step": 34 }, { "epoch": 0.813953488372093, "grad_norm": 0.3651827871799469, "learning_rate": 1.9767441860465116e-05, "loss": 0.6839, "step": 35 }, { "epoch": 0.8372093023255814, "grad_norm": 0.5120950937271118, "learning_rate": 2.0348837209302328e-05, "loss": 0.6746, "step": 36 }, { "epoch": 0.8604651162790697, "grad_norm": 0.4111502766609192, "learning_rate": 2.0930232558139536e-05, "loss": 0.6805, "step": 37 }, { "epoch": 0.8837209302325582, "grad_norm": 0.44530022144317627, "learning_rate": 2.1511627906976744e-05, "loss": 0.6833, "step": 38 }, { "epoch": 0.9069767441860465, "grad_norm": 0.4006333649158478, "learning_rate": 2.2093023255813955e-05, "loss": 0.6769, "step": 39 }, { "epoch": 0.9302325581395349, "grad_norm": 0.4231868386268616, "learning_rate": 2.2674418604651163e-05, "loss": 0.6721, "step": 40 }, { "epoch": 0.9302325581395349, "eval_accuracy": 0.659016393442623, "eval_f1": 0.649208174820844, "eval_loss": 0.6683934926986694, "eval_precision": 0.6947368421052631, "eval_recall": 0.6673559059188421, "eval_runtime": 0.0782, "eval_samples_per_second": 3899.573, "eval_steps_per_second": 63.927, "step": 40 }, { "epoch": 0.9534883720930233, "grad_norm": 0.4570966064929962, "learning_rate": 2.3255813953488374e-05, "loss": 0.679, "step": 41 }, { "epoch": 0.9767441860465116, "grad_norm": 0.6141464710235596, "learning_rate": 2.3837209302325582e-05, "loss": 0.6768, "step": 42 }, { "epoch": 1.0, "grad_norm": 0.6506465673446655, "learning_rate": 2.441860465116279e-05, "loss": 0.6573, "step": 43 }, { "epoch": 1.0232558139534884, "grad_norm": 1.0706639289855957, "learning_rate": 2.5e-05, "loss": 0.6865, "step": 44 }, { "epoch": 1.0465116279069768, "grad_norm": 0.4623822867870331, "learning_rate": 2.5581395348837212e-05, "loss": 0.6538, "step": 45 }, { "epoch": 1.069767441860465, "grad_norm": 0.4986269176006317, "learning_rate": 2.616279069767442e-05, "loss": 0.6585, "step": 46 }, { "epoch": 1.0930232558139534, "grad_norm": 0.48053082823753357, "learning_rate": 2.674418604651163e-05, "loss": 0.6506, "step": 47 }, { "epoch": 1.1162790697674418, "grad_norm": 0.5987160801887512, "learning_rate": 2.7325581395348836e-05, "loss": 0.6691, "step": 48 }, { "epoch": 1.1395348837209303, "grad_norm": 0.685001015663147, "learning_rate": 2.7906976744186048e-05, "loss": 0.6512, "step": 49 }, { "epoch": 1.1627906976744187, "grad_norm": 0.5840403437614441, "learning_rate": 2.848837209302326e-05, "loss": 0.6759, "step": 50 }, { "epoch": 1.1860465116279069, "grad_norm": 0.6035524606704712, "learning_rate": 2.9069767441860467e-05, "loss": 0.6502, "step": 51 }, { "epoch": 1.2093023255813953, "grad_norm": 0.6236196160316467, "learning_rate": 2.9651162790697678e-05, "loss": 0.6575, "step": 52 }, { "epoch": 1.2325581395348837, "grad_norm": 0.5840880274772644, "learning_rate": 3.0232558139534883e-05, "loss": 0.6438, "step": 53 }, { "epoch": 1.255813953488372, "grad_norm": 0.7563631534576416, "learning_rate": 3.081395348837209e-05, "loss": 0.645, "step": 54 }, { "epoch": 1.2790697674418605, "grad_norm": 0.8720300793647766, "learning_rate": 3.13953488372093e-05, "loss": 0.6551, "step": 55 }, { "epoch": 1.302325581395349, "grad_norm": 0.5763499140739441, "learning_rate": 3.197674418604651e-05, "loss": 0.6304, "step": 56 }, { "epoch": 1.3255813953488373, "grad_norm": 0.5487861037254333, "learning_rate": 3.2558139534883724e-05, "loss": 0.6373, "step": 57 }, { "epoch": 1.3488372093023255, "grad_norm": 0.5714741945266724, "learning_rate": 3.313953488372093e-05, "loss": 0.638, "step": 58 }, { "epoch": 1.372093023255814, "grad_norm": 0.7604700326919556, "learning_rate": 3.372093023255814e-05, "loss": 0.6537, "step": 59 }, { "epoch": 1.3953488372093024, "grad_norm": 0.8128901720046997, "learning_rate": 3.430232558139535e-05, "loss": 0.6573, "step": 60 }, { "epoch": 1.3953488372093024, "eval_accuracy": 0.760655737704918, "eval_f1": 0.7589037239168805, "eval_loss": 0.6135660409927368, "eval_precision": 0.7778153153153153, "eval_recall": 0.7656801929869905, "eval_runtime": 0.0732, "eval_samples_per_second": 4169.193, "eval_steps_per_second": 68.347, "step": 60 }, { "epoch": 1.4186046511627908, "grad_norm": 0.6136202216148376, "learning_rate": 3.488372093023256e-05, "loss": 0.6257, "step": 61 }, { "epoch": 1.441860465116279, "grad_norm": 0.6842197179794312, "learning_rate": 3.5465116279069774e-05, "loss": 0.6065, "step": 62 }, { "epoch": 1.4651162790697674, "grad_norm": 0.6682329177856445, "learning_rate": 3.604651162790698e-05, "loss": 0.6114, "step": 63 }, { "epoch": 1.4883720930232558, "grad_norm": 0.7913668155670166, "learning_rate": 3.662790697674418e-05, "loss": 0.636, "step": 64 }, { "epoch": 1.5116279069767442, "grad_norm": 0.7657461762428284, "learning_rate": 3.7209302325581394e-05, "loss": 0.6067, "step": 65 }, { "epoch": 1.5348837209302326, "grad_norm": 0.9346985816955566, "learning_rate": 3.7790697674418606e-05, "loss": 0.5902, "step": 66 }, { "epoch": 1.558139534883721, "grad_norm": 0.7872809767723083, "learning_rate": 3.837209302325582e-05, "loss": 0.5731, "step": 67 }, { "epoch": 1.5813953488372094, "grad_norm": 0.8864495754241943, "learning_rate": 3.895348837209303e-05, "loss": 0.581, "step": 68 }, { "epoch": 1.6046511627906976, "grad_norm": 0.8121451139450073, "learning_rate": 3.953488372093023e-05, "loss": 0.5734, "step": 69 }, { "epoch": 1.627906976744186, "grad_norm": 1.0192168951034546, "learning_rate": 4.0116279069767444e-05, "loss": 0.5645, "step": 70 }, { "epoch": 1.6511627906976745, "grad_norm": 0.9971368312835693, "learning_rate": 4.0697674418604655e-05, "loss": 0.5948, "step": 71 }, { "epoch": 1.6744186046511627, "grad_norm": 0.9530318975448608, "learning_rate": 4.127906976744187e-05, "loss": 0.5129, "step": 72 }, { "epoch": 1.697674418604651, "grad_norm": 0.980096161365509, "learning_rate": 4.186046511627907e-05, "loss": 0.5767, "step": 73 }, { "epoch": 1.7209302325581395, "grad_norm": 0.991642951965332, "learning_rate": 4.2441860465116276e-05, "loss": 0.5403, "step": 74 }, { "epoch": 1.744186046511628, "grad_norm": 0.8262429237365723, "learning_rate": 4.302325581395349e-05, "loss": 0.5467, "step": 75 }, { "epoch": 1.7674418604651163, "grad_norm": 1.1981412172317505, "learning_rate": 4.36046511627907e-05, "loss": 0.5081, "step": 76 }, { "epoch": 1.7906976744186047, "grad_norm": 0.9658230543136597, "learning_rate": 4.418604651162791e-05, "loss": 0.5647, "step": 77 }, { "epoch": 1.8139534883720931, "grad_norm": 1.1805201768875122, "learning_rate": 4.476744186046512e-05, "loss": 0.4536, "step": 78 }, { "epoch": 1.8372093023255816, "grad_norm": 1.101920485496521, "learning_rate": 4.5348837209302326e-05, "loss": 0.5361, "step": 79 }, { "epoch": 1.8604651162790697, "grad_norm": 1.0350321531295776, "learning_rate": 4.593023255813954e-05, "loss": 0.5207, "step": 80 }, { "epoch": 1.8604651162790697, "eval_accuracy": 0.8032786885245902, "eval_f1": 0.8030224748127099, "eval_loss": 0.4702507555484772, "eval_precision": 0.8029135823721811, "eval_recall": 0.8032006547772896, "eval_runtime": 0.0743, "eval_samples_per_second": 4102.555, "eval_steps_per_second": 67.255, "step": 80 }, { "epoch": 1.8837209302325582, "grad_norm": 1.0638865232467651, "learning_rate": 4.651162790697675e-05, "loss": 0.5351, "step": 81 }, { "epoch": 1.9069767441860463, "grad_norm": 1.7043132781982422, "learning_rate": 4.709302325581396e-05, "loss": 0.4186, "step": 82 }, { "epoch": 1.9302325581395348, "grad_norm": 1.0048333406448364, "learning_rate": 4.7674418604651164e-05, "loss": 0.5134, "step": 83 }, { "epoch": 1.9534883720930232, "grad_norm": 1.5068817138671875, "learning_rate": 4.8255813953488375e-05, "loss": 0.4723, "step": 84 }, { "epoch": 1.9767441860465116, "grad_norm": 1.1574243307113647, "learning_rate": 4.883720930232558e-05, "loss": 0.4904, "step": 85 }, { "epoch": 2.0, "grad_norm": 1.2925702333450317, "learning_rate": 4.941860465116279e-05, "loss": 0.5023, "step": 86 }, { "epoch": 2.0232558139534884, "grad_norm": 1.679551124572754, "learning_rate": 5e-05, "loss": 0.4305, "step": 87 }, { "epoch": 2.046511627906977, "grad_norm": 1.1237919330596924, "learning_rate": 5.0581395348837214e-05, "loss": 0.475, "step": 88 }, { "epoch": 2.0697674418604652, "grad_norm": 1.6137198209762573, "learning_rate": 5.1162790697674425e-05, "loss": 0.4115, "step": 89 }, { "epoch": 2.0930232558139537, "grad_norm": 1.814559817314148, "learning_rate": 5.1744186046511636e-05, "loss": 0.4894, "step": 90 }, { "epoch": 2.116279069767442, "grad_norm": 2.1933960914611816, "learning_rate": 5.232558139534884e-05, "loss": 0.3742, "step": 91 }, { "epoch": 2.13953488372093, "grad_norm": 1.3443443775177002, "learning_rate": 5.290697674418605e-05, "loss": 0.396, "step": 92 }, { "epoch": 2.1627906976744184, "grad_norm": 1.5833871364593506, "learning_rate": 5.348837209302326e-05, "loss": 0.3518, "step": 93 }, { "epoch": 2.186046511627907, "grad_norm": 1.5842630863189697, "learning_rate": 5.406976744186046e-05, "loss": 0.4174, "step": 94 }, { "epoch": 2.2093023255813953, "grad_norm": 1.2995549440383911, "learning_rate": 5.465116279069767e-05, "loss": 0.3873, "step": 95 }, { "epoch": 2.2325581395348837, "grad_norm": 2.144974946975708, "learning_rate": 5.5232558139534884e-05, "loss": 0.4805, "step": 96 }, { "epoch": 2.255813953488372, "grad_norm": 1.4312434196472168, "learning_rate": 5.5813953488372095e-05, "loss": 0.4575, "step": 97 }, { "epoch": 2.2790697674418605, "grad_norm": 2.25107741355896, "learning_rate": 5.6395348837209306e-05, "loss": 0.4139, "step": 98 }, { "epoch": 2.302325581395349, "grad_norm": 1.6673591136932373, "learning_rate": 5.697674418604652e-05, "loss": 0.4151, "step": 99 }, { "epoch": 2.3255813953488373, "grad_norm": 1.3179768323898315, "learning_rate": 5.755813953488373e-05, "loss": 0.2997, "step": 100 }, { "epoch": 2.3255813953488373, "eval_accuracy": 0.839344262295082, "eval_f1": 0.839344262295082, "eval_loss": 0.3802424669265747, "eval_precision": 0.8408718876540019, "eval_recall": 0.8408718876540019, "eval_runtime": 0.0744, "eval_samples_per_second": 4101.555, "eval_steps_per_second": 67.239, "step": 100 }, { "epoch": 2.3488372093023258, "grad_norm": 2.245741844177246, "learning_rate": 5.8139534883720933e-05, "loss": 0.3442, "step": 101 }, { "epoch": 2.3720930232558137, "grad_norm": 2.9141011238098145, "learning_rate": 5.8720930232558145e-05, "loss": 0.3489, "step": 102 }, { "epoch": 2.395348837209302, "grad_norm": 1.829503059387207, "learning_rate": 5.9302325581395356e-05, "loss": 0.3544, "step": 103 }, { "epoch": 2.4186046511627906, "grad_norm": 2.233344793319702, "learning_rate": 5.9883720930232554e-05, "loss": 0.3008, "step": 104 }, { "epoch": 2.441860465116279, "grad_norm": 2.4595718383789062, "learning_rate": 6.0465116279069765e-05, "loss": 0.3853, "step": 105 }, { "epoch": 2.4651162790697674, "grad_norm": 1.539181113243103, "learning_rate": 6.104651162790698e-05, "loss": 0.3165, "step": 106 }, { "epoch": 2.488372093023256, "grad_norm": 1.7545955181121826, "learning_rate": 6.162790697674418e-05, "loss": 0.3907, "step": 107 }, { "epoch": 2.511627906976744, "grad_norm": 3.285968065261841, "learning_rate": 6.22093023255814e-05, "loss": 0.4645, "step": 108 }, { "epoch": 2.5348837209302326, "grad_norm": 4.625302314758301, "learning_rate": 6.27906976744186e-05, "loss": 0.2975, "step": 109 }, { "epoch": 2.558139534883721, "grad_norm": 1.9037487506866455, "learning_rate": 6.337209302325582e-05, "loss": 0.3629, "step": 110 }, { "epoch": 2.5813953488372094, "grad_norm": 2.768249988555908, "learning_rate": 6.395348837209303e-05, "loss": 0.385, "step": 111 }, { "epoch": 2.604651162790698, "grad_norm": 1.8411540985107422, "learning_rate": 6.453488372093024e-05, "loss": 0.2562, "step": 112 }, { "epoch": 2.6279069767441863, "grad_norm": 2.3537325859069824, "learning_rate": 6.511627906976745e-05, "loss": 0.2486, "step": 113 }, { "epoch": 2.6511627906976747, "grad_norm": 4.455109596252441, "learning_rate": 6.569767441860465e-05, "loss": 0.2424, "step": 114 }, { "epoch": 2.6744186046511627, "grad_norm": 3.032252550125122, "learning_rate": 6.627906976744186e-05, "loss": 0.3554, "step": 115 }, { "epoch": 2.697674418604651, "grad_norm": 2.142354965209961, "learning_rate": 6.686046511627908e-05, "loss": 0.3701, "step": 116 }, { "epoch": 2.7209302325581395, "grad_norm": 3.0141141414642334, "learning_rate": 6.744186046511628e-05, "loss": 0.4602, "step": 117 }, { "epoch": 2.744186046511628, "grad_norm": 2.129042625427246, "learning_rate": 6.802325581395348e-05, "loss": 0.4295, "step": 118 }, { "epoch": 2.7674418604651163, "grad_norm": 4.583343029022217, "learning_rate": 6.86046511627907e-05, "loss": 0.3599, "step": 119 }, { "epoch": 2.7906976744186047, "grad_norm": 2.799339532852173, "learning_rate": 6.918604651162791e-05, "loss": 0.4383, "step": 120 }, { "epoch": 2.7906976744186047, "eval_accuracy": 0.8426229508196721, "eval_f1": 0.8426077234024254, "eval_loss": 0.30933016538619995, "eval_precision": 0.8432928927895371, "eval_recall": 0.8437365382958559, "eval_runtime": 0.0741, "eval_samples_per_second": 4116.681, "eval_steps_per_second": 67.487, "step": 120 }, { "epoch": 2.813953488372093, "grad_norm": 3.329214096069336, "learning_rate": 6.976744186046513e-05, "loss": 0.2955, "step": 121 }, { "epoch": 2.8372093023255816, "grad_norm": 1.7641863822937012, "learning_rate": 7.034883720930233e-05, "loss": 0.372, "step": 122 }, { "epoch": 2.8604651162790695, "grad_norm": 2.8948066234588623, "learning_rate": 7.093023255813955e-05, "loss": 0.3738, "step": 123 }, { "epoch": 2.883720930232558, "grad_norm": 3.2698655128479004, "learning_rate": 7.151162790697675e-05, "loss": 0.4183, "step": 124 }, { "epoch": 2.9069767441860463, "grad_norm": 2.3984827995300293, "learning_rate": 7.209302325581396e-05, "loss": 0.3729, "step": 125 }, { "epoch": 2.9302325581395348, "grad_norm": 2.4366798400878906, "learning_rate": 7.267441860465116e-05, "loss": 0.291, "step": 126 }, { "epoch": 2.953488372093023, "grad_norm": 1.6613179445266724, "learning_rate": 7.325581395348837e-05, "loss": 0.2673, "step": 127 }, { "epoch": 2.9767441860465116, "grad_norm": 3.9009366035461426, "learning_rate": 7.383720930232558e-05, "loss": 0.3912, "step": 128 }, { "epoch": 3.0, "grad_norm": 6.0683417320251465, "learning_rate": 7.441860465116279e-05, "loss": 0.3879, "step": 129 }, { "epoch": 3.0232558139534884, "grad_norm": 2.161290407180786, "learning_rate": 7.500000000000001e-05, "loss": 0.2878, "step": 130 }, { "epoch": 3.046511627906977, "grad_norm": 5.449894905090332, "learning_rate": 7.558139534883721e-05, "loss": 0.3326, "step": 131 }, { "epoch": 3.0697674418604652, "grad_norm": 2.6764752864837646, "learning_rate": 7.616279069767443e-05, "loss": 0.2813, "step": 132 }, { "epoch": 3.0930232558139537, "grad_norm": 1.1198005676269531, "learning_rate": 7.674418604651163e-05, "loss": 0.2651, "step": 133 }, { "epoch": 3.116279069767442, "grad_norm": 1.7498178482055664, "learning_rate": 7.732558139534884e-05, "loss": 0.2171, "step": 134 }, { "epoch": 3.13953488372093, "grad_norm": 4.505956649780273, "learning_rate": 7.790697674418606e-05, "loss": 0.2529, "step": 135 }, { "epoch": 3.1627906976744184, "grad_norm": 2.509193181991577, "learning_rate": 7.848837209302326e-05, "loss": 0.2102, "step": 136 }, { "epoch": 3.186046511627907, "grad_norm": 1.7064465284347534, "learning_rate": 7.906976744186047e-05, "loss": 0.2097, "step": 137 }, { "epoch": 3.2093023255813953, "grad_norm": 1.7659178972244263, "learning_rate": 7.965116279069767e-05, "loss": 0.1806, "step": 138 }, { "epoch": 3.2325581395348837, "grad_norm": 3.508608818054199, "learning_rate": 8.023255813953489e-05, "loss": 0.1419, "step": 139 }, { "epoch": 3.255813953488372, "grad_norm": 2.2445175647735596, "learning_rate": 8.081395348837209e-05, "loss": 0.4201, "step": 140 }, { "epoch": 3.255813953488372, "eval_accuracy": 0.8622950819672132, "eval_f1": 0.8622225090345896, "eval_loss": 0.31382298469543457, "eval_precision": 0.8622291021671826, "eval_recall": 0.8628844662703541, "eval_runtime": 0.0729, "eval_samples_per_second": 4184.713, "eval_steps_per_second": 68.602, "step": 140 }, { "epoch": 3.2790697674418605, "grad_norm": 2.019263744354248, "learning_rate": 8.139534883720931e-05, "loss": 0.2023, "step": 141 }, { "epoch": 3.302325581395349, "grad_norm": 1.6362420320510864, "learning_rate": 8.197674418604652e-05, "loss": 0.1729, "step": 142 }, { "epoch": 3.3255813953488373, "grad_norm": 2.1134684085845947, "learning_rate": 8.255813953488373e-05, "loss": 0.2688, "step": 143 }, { "epoch": 3.3488372093023258, "grad_norm": 3.522068738937378, "learning_rate": 8.313953488372094e-05, "loss": 0.199, "step": 144 }, { "epoch": 3.3720930232558137, "grad_norm": 4.994085788726807, "learning_rate": 8.372093023255814e-05, "loss": 0.2641, "step": 145 }, { "epoch": 3.395348837209302, "grad_norm": 2.1998469829559326, "learning_rate": 8.430232558139536e-05, "loss": 0.2183, "step": 146 }, { "epoch": 3.4186046511627906, "grad_norm": 1.9816820621490479, "learning_rate": 8.488372093023255e-05, "loss": 0.2952, "step": 147 }, { "epoch": 3.441860465116279, "grad_norm": 2.8008265495300293, "learning_rate": 8.546511627906977e-05, "loss": 0.2562, "step": 148 }, { "epoch": 3.4651162790697674, "grad_norm": 2.082418918609619, "learning_rate": 8.604651162790697e-05, "loss": 0.2626, "step": 149 }, { "epoch": 3.488372093023256, "grad_norm": 3.3041298389434814, "learning_rate": 8.662790697674419e-05, "loss": 0.2373, "step": 150 }, { "epoch": 3.511627906976744, "grad_norm": 2.62347674369812, "learning_rate": 8.72093023255814e-05, "loss": 0.2749, "step": 151 }, { "epoch": 3.5348837209302326, "grad_norm": 3.0671257972717285, "learning_rate": 8.779069767441861e-05, "loss": 0.3538, "step": 152 }, { "epoch": 3.558139534883721, "grad_norm": 2.737697124481201, "learning_rate": 8.837209302325582e-05, "loss": 0.3537, "step": 153 }, { "epoch": 3.5813953488372094, "grad_norm": 6.468879699707031, "learning_rate": 8.895348837209302e-05, "loss": 0.5041, "step": 154 }, { "epoch": 3.604651162790698, "grad_norm": 3.8154025077819824, "learning_rate": 8.953488372093024e-05, "loss": 0.3206, "step": 155 }, { "epoch": 3.6279069767441863, "grad_norm": 1.7028948068618774, "learning_rate": 9.011627906976745e-05, "loss": 0.3093, "step": 156 }, { "epoch": 3.6511627906976747, "grad_norm": 2.867501735687256, "learning_rate": 9.069767441860465e-05, "loss": 0.2656, "step": 157 }, { "epoch": 3.6744186046511627, "grad_norm": 4.052389621734619, "learning_rate": 9.127906976744186e-05, "loss": 0.4386, "step": 158 }, { "epoch": 3.697674418604651, "grad_norm": 1.909553050994873, "learning_rate": 9.186046511627907e-05, "loss": 0.2907, "step": 159 }, { "epoch": 3.7209302325581395, "grad_norm": 1.4470328092575073, "learning_rate": 9.244186046511628e-05, "loss": 0.1697, "step": 160 }, { "epoch": 3.7209302325581395, "eval_accuracy": 0.8721311475409836, "eval_f1": 0.8721256490469893, "eval_loss": 0.26616957783699036, "eval_precision": 0.8730848683077983, "eval_recall": 0.87343844231929, "eval_runtime": 0.0715, "eval_samples_per_second": 4265.076, "eval_steps_per_second": 69.919, "step": 160 }, { "epoch": 3.744186046511628, "grad_norm": 1.7241488695144653, "learning_rate": 9.30232558139535e-05, "loss": 0.248, "step": 161 }, { "epoch": 3.7674418604651163, "grad_norm": 1.370429277420044, "learning_rate": 9.36046511627907e-05, "loss": 0.26, "step": 162 }, { "epoch": 3.7906976744186047, "grad_norm": 2.8360958099365234, "learning_rate": 9.418604651162792e-05, "loss": 0.3161, "step": 163 }, { "epoch": 3.813953488372093, "grad_norm": 2.840177297592163, "learning_rate": 9.476744186046512e-05, "loss": 0.3643, "step": 164 }, { "epoch": 3.8372093023255816, "grad_norm": 1.9618759155273438, "learning_rate": 9.534883720930233e-05, "loss": 0.3486, "step": 165 }, { "epoch": 3.8604651162790695, "grad_norm": 3.201247215270996, "learning_rate": 9.593023255813955e-05, "loss": 0.293, "step": 166 }, { "epoch": 3.883720930232558, "grad_norm": 3.3960511684417725, "learning_rate": 9.651162790697675e-05, "loss": 0.2797, "step": 167 }, { "epoch": 3.9069767441860463, "grad_norm": 2.5416486263275146, "learning_rate": 9.709302325581396e-05, "loss": 0.2781, "step": 168 }, { "epoch": 3.9302325581395348, "grad_norm": 4.113210678100586, "learning_rate": 9.767441860465116e-05, "loss": 0.2734, "step": 169 }, { "epoch": 3.953488372093023, "grad_norm": 3.773771286010742, "learning_rate": 9.825581395348838e-05, "loss": 0.2775, "step": 170 }, { "epoch": 3.9767441860465116, "grad_norm": 1.3982651233673096, "learning_rate": 9.883720930232558e-05, "loss": 0.1802, "step": 171 }, { "epoch": 4.0, "grad_norm": 1.2582266330718994, "learning_rate": 9.94186046511628e-05, "loss": 0.1667, "step": 172 }, { "epoch": 4.023255813953488, "grad_norm": 6.577229976654053, "learning_rate": 0.0001, "loss": 0.4147, "step": 173 }, { "epoch": 4.046511627906977, "grad_norm": 4.923567295074463, "learning_rate": 9.98546511627907e-05, "loss": 0.1983, "step": 174 }, { "epoch": 4.069767441860465, "grad_norm": 1.9299342632293701, "learning_rate": 9.97093023255814e-05, "loss": 0.2169, "step": 175 }, { "epoch": 4.093023255813954, "grad_norm": 3.2299985885620117, "learning_rate": 9.95639534883721e-05, "loss": 0.1698, "step": 176 }, { "epoch": 4.116279069767442, "grad_norm": 2.034163236618042, "learning_rate": 9.94186046511628e-05, "loss": 0.191, "step": 177 }, { "epoch": 4.1395348837209305, "grad_norm": 4.542312145233154, "learning_rate": 9.927325581395349e-05, "loss": 0.2465, "step": 178 }, { "epoch": 4.162790697674419, "grad_norm": 1.8868296146392822, "learning_rate": 9.912790697674418e-05, "loss": 0.2437, "step": 179 }, { "epoch": 4.186046511627907, "grad_norm": 4.527658939361572, "learning_rate": 9.898255813953488e-05, "loss": 0.2774, "step": 180 }, { "epoch": 4.186046511627907, "eval_accuracy": 0.8852459016393442, "eval_f1": 0.8839686525146466, "eval_loss": 0.3168491721153259, "eval_precision": 0.8926612403786605, "eval_recall": 0.8823770138709399, "eval_runtime": 0.0728, "eval_samples_per_second": 4190.498, "eval_steps_per_second": 68.697, "step": 180 }, { "epoch": 4.209302325581396, "grad_norm": 6.06718635559082, "learning_rate": 9.883720930232558e-05, "loss": 0.267, "step": 181 }, { "epoch": 4.232558139534884, "grad_norm": 8.161118507385254, "learning_rate": 9.869186046511628e-05, "loss": 0.459, "step": 182 }, { "epoch": 4.2558139534883725, "grad_norm": 3.9242475032806396, "learning_rate": 9.854651162790698e-05, "loss": 0.2698, "step": 183 }, { "epoch": 4.27906976744186, "grad_norm": 6.657914161682129, "learning_rate": 9.840116279069768e-05, "loss": 0.2114, "step": 184 }, { "epoch": 4.3023255813953485, "grad_norm": 2.23410964012146, "learning_rate": 9.825581395348838e-05, "loss": 0.1337, "step": 185 }, { "epoch": 4.325581395348837, "grad_norm": 1.6781232357025146, "learning_rate": 9.811046511627908e-05, "loss": 0.1269, "step": 186 }, { "epoch": 4.348837209302325, "grad_norm": 1.3690869808197021, "learning_rate": 9.796511627906976e-05, "loss": 0.1783, "step": 187 }, { "epoch": 4.372093023255814, "grad_norm": 2.0360560417175293, "learning_rate": 9.781976744186046e-05, "loss": 0.281, "step": 188 }, { "epoch": 4.395348837209302, "grad_norm": 3.783336877822876, "learning_rate": 9.767441860465116e-05, "loss": 0.3072, "step": 189 }, { "epoch": 4.4186046511627906, "grad_norm": 3.0387606620788574, "learning_rate": 9.752906976744186e-05, "loss": 0.2682, "step": 190 }, { "epoch": 4.441860465116279, "grad_norm": 1.6739147901535034, "learning_rate": 9.738372093023256e-05, "loss": 0.1597, "step": 191 }, { "epoch": 4.465116279069767, "grad_norm": 4.560539245605469, "learning_rate": 9.723837209302326e-05, "loss": 0.2328, "step": 192 }, { "epoch": 4.488372093023256, "grad_norm": 2.564955711364746, "learning_rate": 9.709302325581396e-05, "loss": 0.2731, "step": 193 }, { "epoch": 4.511627906976744, "grad_norm": 6.892247676849365, "learning_rate": 9.694767441860465e-05, "loss": 0.2358, "step": 194 }, { "epoch": 4.534883720930233, "grad_norm": 2.4123950004577637, "learning_rate": 9.680232558139535e-05, "loss": 0.2267, "step": 195 }, { "epoch": 4.558139534883721, "grad_norm": 2.6707470417022705, "learning_rate": 9.665697674418605e-05, "loss": 0.1415, "step": 196 }, { "epoch": 4.5813953488372094, "grad_norm": 1.8982809782028198, "learning_rate": 9.651162790697675e-05, "loss": 0.2119, "step": 197 }, { "epoch": 4.604651162790698, "grad_norm": 3.8330225944519043, "learning_rate": 9.636627906976745e-05, "loss": 0.152, "step": 198 }, { "epoch": 4.627906976744186, "grad_norm": 2.504080057144165, "learning_rate": 9.622093023255815e-05, "loss": 0.1623, "step": 199 }, { "epoch": 4.651162790697675, "grad_norm": 1.6243199110031128, "learning_rate": 9.607558139534885e-05, "loss": 0.1537, "step": 200 }, { "epoch": 4.651162790697675, "eval_accuracy": 0.8950819672131147, "eval_f1": 0.8942532720811303, "eval_loss": 0.2364664524793625, "eval_precision": 0.8994351024697845, "eval_recall": 0.892930989919876, "eval_runtime": 0.0733, "eval_samples_per_second": 4158.906, "eval_steps_per_second": 68.179, "step": 200 }, { "epoch": 4.674418604651163, "grad_norm": 3.362497091293335, "learning_rate": 9.593023255813955e-05, "loss": 0.2097, "step": 201 }, { "epoch": 4.6976744186046515, "grad_norm": 1.36155366897583, "learning_rate": 9.578488372093024e-05, "loss": 0.134, "step": 202 }, { "epoch": 4.720930232558139, "grad_norm": 2.335468053817749, "learning_rate": 9.563953488372094e-05, "loss": 0.2128, "step": 203 }, { "epoch": 4.7441860465116275, "grad_norm": 2.431711196899414, "learning_rate": 9.549418604651163e-05, "loss": 0.2257, "step": 204 }, { "epoch": 4.767441860465116, "grad_norm": 2.425615072250366, "learning_rate": 9.534883720930233e-05, "loss": 0.1691, "step": 205 }, { "epoch": 4.790697674418604, "grad_norm": 3.3857626914978027, "learning_rate": 9.520348837209303e-05, "loss": 0.2282, "step": 206 }, { "epoch": 4.813953488372093, "grad_norm": 2.252249240875244, "learning_rate": 9.505813953488373e-05, "loss": 0.2333, "step": 207 }, { "epoch": 4.837209302325581, "grad_norm": 4.851521015167236, "learning_rate": 9.491279069767442e-05, "loss": 0.2621, "step": 208 }, { "epoch": 4.8604651162790695, "grad_norm": 1.249811053276062, "learning_rate": 9.476744186046512e-05, "loss": 0.075, "step": 209 }, { "epoch": 4.883720930232558, "grad_norm": 1.6059314012527466, "learning_rate": 9.462209302325582e-05, "loss": 0.1878, "step": 210 }, { "epoch": 4.906976744186046, "grad_norm": 7.576606750488281, "learning_rate": 9.447674418604652e-05, "loss": 0.2258, "step": 211 }, { "epoch": 4.930232558139535, "grad_norm": 2.4628143310546875, "learning_rate": 9.433139534883722e-05, "loss": 0.1463, "step": 212 }, { "epoch": 4.953488372093023, "grad_norm": 4.781926155090332, "learning_rate": 9.418604651162792e-05, "loss": 0.2608, "step": 213 }, { "epoch": 4.976744186046512, "grad_norm": 3.2314810752868652, "learning_rate": 9.40406976744186e-05, "loss": 0.2344, "step": 214 }, { "epoch": 5.0, "grad_norm": 3.6826560497283936, "learning_rate": 9.38953488372093e-05, "loss": 0.17, "step": 215 }, { "epoch": 5.023255813953488, "grad_norm": 1.804560661315918, "learning_rate": 9.375e-05, "loss": 0.1241, "step": 216 }, { "epoch": 5.046511627906977, "grad_norm": 3.6199300289154053, "learning_rate": 9.36046511627907e-05, "loss": 0.1335, "step": 217 }, { "epoch": 5.069767441860465, "grad_norm": 2.37376070022583, "learning_rate": 9.34593023255814e-05, "loss": 0.1883, "step": 218 }, { "epoch": 5.093023255813954, "grad_norm": 4.643337726593018, "learning_rate": 9.33139534883721e-05, "loss": 0.187, "step": 219 }, { "epoch": 5.116279069767442, "grad_norm": 6.206487655639648, "learning_rate": 9.31686046511628e-05, "loss": 0.166, "step": 220 }, { "epoch": 5.116279069767442, "eval_accuracy": 0.8819672131147541, "eval_f1": 0.8815073815073815, "eval_loss": 0.27603965997695923, "eval_precision": 0.8827922077922078, "eval_recall": 0.8809123804600671, "eval_runtime": 0.0725, "eval_samples_per_second": 4205.887, "eval_steps_per_second": 68.949, "step": 220 }, { "epoch": 5.1395348837209305, "grad_norm": 2.4301323890686035, "learning_rate": 9.30232558139535e-05, "loss": 0.0946, "step": 221 }, { "epoch": 5.162790697674419, "grad_norm": 4.870151519775391, "learning_rate": 9.28779069767442e-05, "loss": 0.2674, "step": 222 }, { "epoch": 5.186046511627907, "grad_norm": 5.952517032623291, "learning_rate": 9.273255813953488e-05, "loss": 0.2605, "step": 223 }, { "epoch": 5.209302325581396, "grad_norm": 1.8462685346603394, "learning_rate": 9.258720930232558e-05, "loss": 0.1902, "step": 224 }, { "epoch": 5.232558139534884, "grad_norm": 2.608715295791626, "learning_rate": 9.244186046511628e-05, "loss": 0.1392, "step": 225 }, { "epoch": 5.2558139534883725, "grad_norm": 4.296781539916992, "learning_rate": 9.229651162790698e-05, "loss": 0.2599, "step": 226 }, { "epoch": 5.27906976744186, "grad_norm": 6.4339165687561035, "learning_rate": 9.215116279069768e-05, "loss": 0.162, "step": 227 }, { "epoch": 5.3023255813953485, "grad_norm": 4.432989597320557, "learning_rate": 9.200581395348837e-05, "loss": 0.1832, "step": 228 }, { "epoch": 5.325581395348837, "grad_norm": 1.9655786752700806, "learning_rate": 9.186046511627907e-05, "loss": 0.1192, "step": 229 }, { "epoch": 5.348837209302325, "grad_norm": 4.3318891525268555, "learning_rate": 9.171511627906977e-05, "loss": 0.244, "step": 230 }, { "epoch": 5.372093023255814, "grad_norm": 2.993924856185913, "learning_rate": 9.156976744186047e-05, "loss": 0.1828, "step": 231 }, { "epoch": 5.395348837209302, "grad_norm": 2.020063638687134, "learning_rate": 9.142441860465116e-05, "loss": 0.1642, "step": 232 }, { "epoch": 5.4186046511627906, "grad_norm": 2.384181499481201, "learning_rate": 9.127906976744186e-05, "loss": 0.1752, "step": 233 }, { "epoch": 5.441860465116279, "grad_norm": 1.9194653034210205, "learning_rate": 9.113372093023255e-05, "loss": 0.0996, "step": 234 }, { "epoch": 5.465116279069767, "grad_norm": 3.680722713470459, "learning_rate": 9.098837209302325e-05, "loss": 0.1925, "step": 235 }, { "epoch": 5.488372093023256, "grad_norm": 3.332195520401001, "learning_rate": 9.084302325581395e-05, "loss": 0.13, "step": 236 }, { "epoch": 5.511627906976744, "grad_norm": 2.6394619941711426, "learning_rate": 9.069767441860465e-05, "loss": 0.193, "step": 237 }, { "epoch": 5.534883720930233, "grad_norm": 4.5959038734436035, "learning_rate": 9.055232558139536e-05, "loss": 0.2303, "step": 238 }, { "epoch": 5.558139534883721, "grad_norm": 6.544390678405762, "learning_rate": 9.040697674418606e-05, "loss": 0.2432, "step": 239 }, { "epoch": 5.5813953488372094, "grad_norm": 1.6620129346847534, "learning_rate": 9.026162790697675e-05, "loss": 0.1378, "step": 240 }, { "epoch": 5.5813953488372094, "eval_accuracy": 0.8819672131147541, "eval_f1": 0.8816810344827587, "eval_loss": 0.21341156959533691, "eval_precision": 0.8819660110420979, "eval_recall": 0.8814723873524597, "eval_runtime": 0.091, "eval_samples_per_second": 3353.025, "eval_steps_per_second": 54.968, "step": 240 }, { "epoch": 5.604651162790698, "grad_norm": 1.251617431640625, "learning_rate": 9.011627906976745e-05, "loss": 0.1108, "step": 241 }, { "epoch": 5.627906976744186, "grad_norm": 2.331637144088745, "learning_rate": 8.997093023255815e-05, "loss": 0.1591, "step": 242 }, { "epoch": 5.651162790697675, "grad_norm": 2.126330852508545, "learning_rate": 8.982558139534884e-05, "loss": 0.1115, "step": 243 }, { "epoch": 5.674418604651163, "grad_norm": 2.0312886238098145, "learning_rate": 8.968023255813954e-05, "loss": 0.1009, "step": 244 }, { "epoch": 5.6976744186046515, "grad_norm": 3.0067551136016846, "learning_rate": 8.953488372093024e-05, "loss": 0.1851, "step": 245 }, { "epoch": 5.720930232558139, "grad_norm": 3.1301369667053223, "learning_rate": 8.938953488372094e-05, "loss": 0.107, "step": 246 }, { "epoch": 5.7441860465116275, "grad_norm": 6.043979167938232, "learning_rate": 8.924418604651164e-05, "loss": 0.273, "step": 247 }, { "epoch": 5.767441860465116, "grad_norm": 3.7104833126068115, "learning_rate": 8.909883720930234e-05, "loss": 0.1299, "step": 248 }, { "epoch": 5.790697674418604, "grad_norm": 2.3856964111328125, "learning_rate": 8.895348837209302e-05, "loss": 0.1252, "step": 249 }, { "epoch": 5.813953488372093, "grad_norm": 5.702451705932617, "learning_rate": 8.880813953488372e-05, "loss": 0.1939, "step": 250 }, { "epoch": 5.837209302325581, "grad_norm": 3.5763566493988037, "learning_rate": 8.866279069767442e-05, "loss": 0.2426, "step": 251 }, { "epoch": 5.8604651162790695, "grad_norm": 4.04267692565918, "learning_rate": 8.851744186046512e-05, "loss": 0.1966, "step": 252 }, { "epoch": 5.883720930232558, "grad_norm": 2.391784191131592, "learning_rate": 8.837209302325582e-05, "loss": 0.2284, "step": 253 }, { "epoch": 5.906976744186046, "grad_norm": 4.925626754760742, "learning_rate": 8.822674418604652e-05, "loss": 0.1559, "step": 254 }, { "epoch": 5.930232558139535, "grad_norm": 5.659231185913086, "learning_rate": 8.808139534883722e-05, "loss": 0.1527, "step": 255 }, { "epoch": 5.953488372093023, "grad_norm": 1.3519715070724487, "learning_rate": 8.793604651162792e-05, "loss": 0.097, "step": 256 }, { "epoch": 5.976744186046512, "grad_norm": 3.517763137817383, "learning_rate": 8.779069767441861e-05, "loss": 0.2737, "step": 257 }, { "epoch": 6.0, "grad_norm": 3.0338218212127686, "learning_rate": 8.76453488372093e-05, "loss": 0.1154, "step": 258 }, { "epoch": 6.023255813953488, "grad_norm": 2.716304063796997, "learning_rate": 8.75e-05, "loss": 0.2143, "step": 259 }, { "epoch": 6.046511627906977, "grad_norm": 1.8077433109283447, "learning_rate": 8.73546511627907e-05, "loss": 0.1411, "step": 260 }, { "epoch": 6.046511627906977, "eval_accuracy": 0.8950819672131147, "eval_f1": 0.8945822172297591, "eval_loss": 0.23519852757453918, "eval_precision": 0.8966414996094767, "eval_recall": 0.8937710002584647, "eval_runtime": 0.0746, "eval_samples_per_second": 4089.283, "eval_steps_per_second": 67.037, "step": 260 }, { "epoch": 6.069767441860465, "grad_norm": 4.399131774902344, "learning_rate": 8.72093023255814e-05, "loss": 0.1678, "step": 261 }, { "epoch": 6.093023255813954, "grad_norm": 1.8267767429351807, "learning_rate": 8.70639534883721e-05, "loss": 0.0858, "step": 262 }, { "epoch": 6.116279069767442, "grad_norm": 4.386148929595947, "learning_rate": 8.69186046511628e-05, "loss": 0.1666, "step": 263 }, { "epoch": 6.1395348837209305, "grad_norm": 4.314615249633789, "learning_rate": 8.67732558139535e-05, "loss": 0.1782, "step": 264 }, { "epoch": 6.162790697674419, "grad_norm": 1.9564558267593384, "learning_rate": 8.662790697674419e-05, "loss": 0.203, "step": 265 }, { "epoch": 6.186046511627907, "grad_norm": 3.557976484298706, "learning_rate": 8.648255813953489e-05, "loss": 0.1347, "step": 266 }, { "epoch": 6.209302325581396, "grad_norm": 2.464057207107544, "learning_rate": 8.633720930232559e-05, "loss": 0.0906, "step": 267 }, { "epoch": 6.232558139534884, "grad_norm": 2.893970251083374, "learning_rate": 8.619186046511628e-05, "loss": 0.0893, "step": 268 }, { "epoch": 6.2558139534883725, "grad_norm": 4.821974277496338, "learning_rate": 8.604651162790697e-05, "loss": 0.1008, "step": 269 }, { "epoch": 6.27906976744186, "grad_norm": 2.767670154571533, "learning_rate": 8.590116279069767e-05, "loss": 0.1072, "step": 270 }, { "epoch": 6.3023255813953485, "grad_norm": 2.337888240814209, "learning_rate": 8.575581395348837e-05, "loss": 0.1691, "step": 271 }, { "epoch": 6.325581395348837, "grad_norm": 2.4074652194976807, "learning_rate": 8.561046511627907e-05, "loss": 0.1671, "step": 272 }, { "epoch": 6.348837209302325, "grad_norm": 6.366968631744385, "learning_rate": 8.546511627906977e-05, "loss": 0.1741, "step": 273 }, { "epoch": 6.372093023255814, "grad_norm": 2.4231204986572266, "learning_rate": 8.531976744186047e-05, "loss": 0.1563, "step": 274 }, { "epoch": 6.395348837209302, "grad_norm": 3.424386739730835, "learning_rate": 8.517441860465117e-05, "loss": 0.2211, "step": 275 }, { "epoch": 6.4186046511627906, "grad_norm": 6.0630717277526855, "learning_rate": 8.502906976744187e-05, "loss": 0.3506, "step": 276 }, { "epoch": 6.441860465116279, "grad_norm": 2.5241973400115967, "learning_rate": 8.488372093023255e-05, "loss": 0.1239, "step": 277 }, { "epoch": 6.465116279069767, "grad_norm": 1.1857047080993652, "learning_rate": 8.473837209302325e-05, "loss": 0.1218, "step": 278 }, { "epoch": 6.488372093023256, "grad_norm": 2.826883316040039, "learning_rate": 8.459302325581395e-05, "loss": 0.1296, "step": 279 }, { "epoch": 6.511627906976744, "grad_norm": 3.720126152038574, "learning_rate": 8.444767441860465e-05, "loss": 0.1594, "step": 280 }, { "epoch": 6.511627906976744, "eval_accuracy": 0.8819672131147541, "eval_f1": 0.8815073815073815, "eval_loss": 0.25715914368629456, "eval_precision": 0.8827922077922078, "eval_recall": 0.8809123804600671, "eval_runtime": 0.0733, "eval_samples_per_second": 4160.082, "eval_steps_per_second": 68.198, "step": 280 }, { "epoch": 6.534883720930233, "grad_norm": 2.7126853466033936, "learning_rate": 8.430232558139536e-05, "loss": 0.1234, "step": 281 }, { "epoch": 6.558139534883721, "grad_norm": 2.641782522201538, "learning_rate": 8.415697674418606e-05, "loss": 0.1515, "step": 282 }, { "epoch": 6.5813953488372094, "grad_norm": 5.715346813201904, "learning_rate": 8.401162790697676e-05, "loss": 0.1463, "step": 283 }, { "epoch": 6.604651162790698, "grad_norm": 4.03624153137207, "learning_rate": 8.386627906976746e-05, "loss": 0.1425, "step": 284 }, { "epoch": 6.627906976744186, "grad_norm": 0.8614564538002014, "learning_rate": 8.372093023255814e-05, "loss": 0.1115, "step": 285 }, { "epoch": 6.651162790697675, "grad_norm": 3.0801596641540527, "learning_rate": 8.357558139534884e-05, "loss": 0.1954, "step": 286 }, { "epoch": 6.674418604651163, "grad_norm": 2.3671586513519287, "learning_rate": 8.343023255813954e-05, "loss": 0.1223, "step": 287 }, { "epoch": 6.6976744186046515, "grad_norm": 4.468645095825195, "learning_rate": 8.328488372093024e-05, "loss": 0.1422, "step": 288 }, { "epoch": 6.720930232558139, "grad_norm": 2.540923833847046, "learning_rate": 8.313953488372094e-05, "loss": 0.127, "step": 289 }, { "epoch": 6.7441860465116275, "grad_norm": 1.3170732259750366, "learning_rate": 8.299418604651164e-05, "loss": 0.0479, "step": 290 }, { "epoch": 6.767441860465116, "grad_norm": 3.5731353759765625, "learning_rate": 8.284883720930234e-05, "loss": 0.1118, "step": 291 }, { "epoch": 6.790697674418604, "grad_norm": 6.928558826446533, "learning_rate": 8.270348837209303e-05, "loss": 0.2011, "step": 292 }, { "epoch": 6.813953488372093, "grad_norm": 4.524282932281494, "learning_rate": 8.255813953488373e-05, "loss": 0.1252, "step": 293 }, { "epoch": 6.837209302325581, "grad_norm": 7.167698383331299, "learning_rate": 8.241279069767442e-05, "loss": 0.153, "step": 294 }, { "epoch": 6.8604651162790695, "grad_norm": 4.981342792510986, "learning_rate": 8.226744186046512e-05, "loss": 0.2423, "step": 295 }, { "epoch": 6.883720930232558, "grad_norm": 2.139774799346924, "learning_rate": 8.212209302325582e-05, "loss": 0.1548, "step": 296 }, { "epoch": 6.906976744186046, "grad_norm": 2.5026113986968994, "learning_rate": 8.197674418604652e-05, "loss": 0.0935, "step": 297 }, { "epoch": 6.930232558139535, "grad_norm": 1.633834719657898, "learning_rate": 8.183139534883721e-05, "loss": 0.1224, "step": 298 }, { "epoch": 6.953488372093023, "grad_norm": 4.0517168045043945, "learning_rate": 8.168604651162791e-05, "loss": 0.106, "step": 299 }, { "epoch": 6.976744186046512, "grad_norm": 4.622942924499512, "learning_rate": 8.154069767441861e-05, "loss": 0.1923, "step": 300 }, { "epoch": 6.976744186046512, "eval_accuracy": 0.8918032786885246, "eval_f1": 0.8912373974778747, "eval_loss": 0.276712030172348, "eval_precision": 0.8937043795620438, "eval_recall": 0.8903463427242182, "eval_runtime": 0.0737, "eval_samples_per_second": 4140.129, "eval_steps_per_second": 67.871, "step": 300 }, { "epoch": 7.0, "grad_norm": 3.300420045852661, "learning_rate": 8.139534883720931e-05, "loss": 0.1218, "step": 301 }, { "epoch": 7.023255813953488, "grad_norm": 2.9694247245788574, "learning_rate": 8.125000000000001e-05, "loss": 0.1067, "step": 302 }, { "epoch": 7.046511627906977, "grad_norm": 3.805917978286743, "learning_rate": 8.11046511627907e-05, "loss": 0.1165, "step": 303 }, { "epoch": 7.069767441860465, "grad_norm": 2.0280096530914307, "learning_rate": 8.09593023255814e-05, "loss": 0.119, "step": 304 }, { "epoch": 7.093023255813954, "grad_norm": 5.049257278442383, "learning_rate": 8.081395348837209e-05, "loss": 0.1838, "step": 305 }, { "epoch": 7.116279069767442, "grad_norm": 6.135448455810547, "learning_rate": 8.066860465116279e-05, "loss": 0.2193, "step": 306 }, { "epoch": 7.1395348837209305, "grad_norm": 2.0232150554656982, "learning_rate": 8.052325581395349e-05, "loss": 0.0829, "step": 307 }, { "epoch": 7.162790697674419, "grad_norm": 3.0096096992492676, "learning_rate": 8.037790697674419e-05, "loss": 0.1394, "step": 308 }, { "epoch": 7.186046511627907, "grad_norm": 0.9171739220619202, "learning_rate": 8.023255813953489e-05, "loss": 0.1058, "step": 309 }, { "epoch": 7.209302325581396, "grad_norm": 3.929137706756592, "learning_rate": 8.008720930232559e-05, "loss": 0.2189, "step": 310 }, { "epoch": 7.232558139534884, "grad_norm": 1.1936514377593994, "learning_rate": 7.994186046511629e-05, "loss": 0.0981, "step": 311 }, { "epoch": 7.2558139534883725, "grad_norm": 2.542099714279175, "learning_rate": 7.979651162790697e-05, "loss": 0.1052, "step": 312 }, { "epoch": 7.27906976744186, "grad_norm": 2.42411732673645, "learning_rate": 7.965116279069767e-05, "loss": 0.0834, "step": 313 }, { "epoch": 7.3023255813953485, "grad_norm": 1.07964289188385, "learning_rate": 7.950581395348837e-05, "loss": 0.0684, "step": 314 }, { "epoch": 7.325581395348837, "grad_norm": 0.9596773386001587, "learning_rate": 7.936046511627907e-05, "loss": 0.0791, "step": 315 }, { "epoch": 7.348837209302325, "grad_norm": 3.508575201034546, "learning_rate": 7.921511627906977e-05, "loss": 0.145, "step": 316 }, { "epoch": 7.372093023255814, "grad_norm": 3.8409645557403564, "learning_rate": 7.906976744186047e-05, "loss": 0.1057, "step": 317 }, { "epoch": 7.395348837209302, "grad_norm": 5.247814655303955, "learning_rate": 7.892441860465116e-05, "loss": 0.0731, "step": 318 }, { "epoch": 7.4186046511627906, "grad_norm": 2.5410549640655518, "learning_rate": 7.877906976744186e-05, "loss": 0.2125, "step": 319 }, { "epoch": 7.441860465116279, "grad_norm": 3.2063095569610596, "learning_rate": 7.863372093023256e-05, "loss": 0.0896, "step": 320 }, { "epoch": 7.441860465116279, "eval_accuracy": 0.9114754098360656, "eval_f1": 0.9114411381991419, "eval_loss": 0.22706902027130127, "eval_precision": 0.9115497076023391, "eval_recall": 0.9122943051606789, "eval_runtime": 0.0744, "eval_samples_per_second": 4097.456, "eval_steps_per_second": 67.171, "step": 320 }, { "epoch": 7.465116279069767, "grad_norm": 1.1547659635543823, "learning_rate": 7.848837209302326e-05, "loss": 0.0425, "step": 321 }, { "epoch": 7.488372093023256, "grad_norm": 10.542337417602539, "learning_rate": 7.834302325581395e-05, "loss": 0.4056, "step": 322 }, { "epoch": 7.511627906976744, "grad_norm": 6.774454593658447, "learning_rate": 7.819767441860465e-05, "loss": 0.106, "step": 323 }, { "epoch": 7.534883720930233, "grad_norm": 3.680190086364746, "learning_rate": 7.805232558139536e-05, "loss": 0.1208, "step": 324 }, { "epoch": 7.558139534883721, "grad_norm": 3.1805880069732666, "learning_rate": 7.790697674418606e-05, "loss": 0.0968, "step": 325 }, { "epoch": 7.5813953488372094, "grad_norm": 3.887376070022583, "learning_rate": 7.776162790697676e-05, "loss": 0.3548, "step": 326 }, { "epoch": 7.604651162790698, "grad_norm": 2.4842185974121094, "learning_rate": 7.761627906976745e-05, "loss": 0.0782, "step": 327 }, { "epoch": 7.627906976744186, "grad_norm": 4.861859321594238, "learning_rate": 7.747093023255815e-05, "loss": 0.1529, "step": 328 }, { "epoch": 7.651162790697675, "grad_norm": 4.307983875274658, "learning_rate": 7.732558139534884e-05, "loss": 0.1938, "step": 329 }, { "epoch": 7.674418604651163, "grad_norm": 6.8119587898254395, "learning_rate": 7.718023255813954e-05, "loss": 0.1118, "step": 330 }, { "epoch": 7.6976744186046515, "grad_norm": 1.8230775594711304, "learning_rate": 7.703488372093024e-05, "loss": 0.0551, "step": 331 }, { "epoch": 7.720930232558139, "grad_norm": 1.6934735774993896, "learning_rate": 7.688953488372094e-05, "loss": 0.0629, "step": 332 }, { "epoch": 7.7441860465116275, "grad_norm": 2.537760019302368, "learning_rate": 7.674418604651163e-05, "loss": 0.1084, "step": 333 }, { "epoch": 7.767441860465116, "grad_norm": 1.6460330486297607, "learning_rate": 7.659883720930233e-05, "loss": 0.0927, "step": 334 }, { "epoch": 7.790697674418604, "grad_norm": 1.8621909618377686, "learning_rate": 7.645348837209303e-05, "loss": 0.107, "step": 335 }, { "epoch": 7.813953488372093, "grad_norm": 7.628468036651611, "learning_rate": 7.630813953488373e-05, "loss": 0.1038, "step": 336 }, { "epoch": 7.837209302325581, "grad_norm": 2.3361170291900635, "learning_rate": 7.616279069767443e-05, "loss": 0.1017, "step": 337 }, { "epoch": 7.8604651162790695, "grad_norm": 1.8995561599731445, "learning_rate": 7.601744186046513e-05, "loss": 0.0414, "step": 338 }, { "epoch": 7.883720930232558, "grad_norm": 5.01551628112793, "learning_rate": 7.587209302325581e-05, "loss": 0.1229, "step": 339 }, { "epoch": 7.906976744186046, "grad_norm": 4.123791694641113, "learning_rate": 7.572674418604651e-05, "loss": 0.099, "step": 340 }, { "epoch": 7.906976744186046, "eval_accuracy": 0.9016393442622951, "eval_f1": 0.9016129032258065, "eval_loss": 0.22697412967681885, "eval_precision": 0.9018878472520857, "eval_recall": 0.9025803394503318, "eval_runtime": 0.0787, "eval_samples_per_second": 3874.417, "eval_steps_per_second": 63.515, "step": 340 }, { "epoch": 7.930232558139535, "grad_norm": 1.9877846240997314, "learning_rate": 7.558139534883721e-05, "loss": 0.0625, "step": 341 }, { "epoch": 7.953488372093023, "grad_norm": 1.6333461999893188, "learning_rate": 7.543604651162791e-05, "loss": 0.051, "step": 342 }, { "epoch": 7.976744186046512, "grad_norm": 8.968517303466797, "learning_rate": 7.529069767441861e-05, "loss": 0.0781, "step": 343 }, { "epoch": 8.0, "grad_norm": 2.306673765182495, "learning_rate": 7.514534883720931e-05, "loss": 0.0684, "step": 344 }, { "epoch": 8.023255813953488, "grad_norm": 5.471800804138184, "learning_rate": 7.500000000000001e-05, "loss": 0.1, "step": 345 }, { "epoch": 8.046511627906977, "grad_norm": 6.212913513183594, "learning_rate": 7.48546511627907e-05, "loss": 0.0966, "step": 346 }, { "epoch": 8.069767441860465, "grad_norm": 7.509064197540283, "learning_rate": 7.47093023255814e-05, "loss": 0.2218, "step": 347 }, { "epoch": 8.093023255813954, "grad_norm": 5.755818843841553, "learning_rate": 7.456395348837209e-05, "loss": 0.2037, "step": 348 }, { "epoch": 8.116279069767442, "grad_norm": 2.3190879821777344, "learning_rate": 7.441860465116279e-05, "loss": 0.1259, "step": 349 }, { "epoch": 8.13953488372093, "grad_norm": 0.9036604166030884, "learning_rate": 7.427325581395349e-05, "loss": 0.0662, "step": 350 }, { "epoch": 8.162790697674419, "grad_norm": 2.2601730823516846, "learning_rate": 7.412790697674419e-05, "loss": 0.0833, "step": 351 }, { "epoch": 8.186046511627907, "grad_norm": 2.334266185760498, "learning_rate": 7.398255813953489e-05, "loss": 0.0482, "step": 352 }, { "epoch": 8.209302325581396, "grad_norm": 5.374472141265869, "learning_rate": 7.383720930232558e-05, "loss": 0.1771, "step": 353 }, { "epoch": 8.232558139534884, "grad_norm": 5.360229969024658, "learning_rate": 7.369186046511628e-05, "loss": 0.2253, "step": 354 }, { "epoch": 8.255813953488373, "grad_norm": 2.7183334827423096, "learning_rate": 7.354651162790698e-05, "loss": 0.13, "step": 355 }, { "epoch": 8.279069767441861, "grad_norm": 1.2531534433364868, "learning_rate": 7.340116279069768e-05, "loss": 0.0898, "step": 356 }, { "epoch": 8.30232558139535, "grad_norm": 5.129657745361328, "learning_rate": 7.325581395348837e-05, "loss": 0.0884, "step": 357 }, { "epoch": 8.325581395348838, "grad_norm": 0.5411895513534546, "learning_rate": 7.311046511627907e-05, "loss": 0.0227, "step": 358 }, { "epoch": 8.348837209302326, "grad_norm": 3.1815500259399414, "learning_rate": 7.296511627906976e-05, "loss": 0.1153, "step": 359 }, { "epoch": 8.372093023255815, "grad_norm": 3.956552743911743, "learning_rate": 7.281976744186046e-05, "loss": 0.1048, "step": 360 }, { "epoch": 8.372093023255815, "eval_accuracy": 0.8918032786885246, "eval_f1": 0.8917846660000646, "eval_loss": 0.2492106705904007, "eval_precision": 0.8922580645161291, "eval_recall": 0.8928663737399845, "eval_runtime": 0.0761, "eval_samples_per_second": 4010.404, "eval_steps_per_second": 65.744, "step": 360 }, { "epoch": 8.395348837209303, "grad_norm": 3.400632619857788, "learning_rate": 7.267441860465116e-05, "loss": 0.1008, "step": 361 }, { "epoch": 8.418604651162791, "grad_norm": 4.204819202423096, "learning_rate": 7.252906976744186e-05, "loss": 0.0525, "step": 362 }, { "epoch": 8.44186046511628, "grad_norm": 4.265756130218506, "learning_rate": 7.238372093023256e-05, "loss": 0.1789, "step": 363 }, { "epoch": 8.465116279069768, "grad_norm": 3.7337512969970703, "learning_rate": 7.223837209302326e-05, "loss": 0.2187, "step": 364 }, { "epoch": 8.488372093023255, "grad_norm": 7.506768226623535, "learning_rate": 7.209302325581396e-05, "loss": 0.1643, "step": 365 }, { "epoch": 8.511627906976745, "grad_norm": 2.040639877319336, "learning_rate": 7.194767441860464e-05, "loss": 0.0645, "step": 366 }, { "epoch": 8.534883720930232, "grad_norm": 3.809933662414551, "learning_rate": 7.180232558139535e-05, "loss": 0.1413, "step": 367 }, { "epoch": 8.55813953488372, "grad_norm": 4.585285186767578, "learning_rate": 7.165697674418605e-05, "loss": 0.073, "step": 368 }, { "epoch": 8.581395348837209, "grad_norm": 1.4243372678756714, "learning_rate": 7.151162790697675e-05, "loss": 0.0358, "step": 369 }, { "epoch": 8.604651162790697, "grad_norm": 7.1119513511657715, "learning_rate": 7.136627906976745e-05, "loss": 0.1204, "step": 370 }, { "epoch": 8.627906976744185, "grad_norm": 4.922991752624512, "learning_rate": 7.122093023255815e-05, "loss": 0.1183, "step": 371 }, { "epoch": 8.651162790697674, "grad_norm": 2.3582427501678467, "learning_rate": 7.107558139534885e-05, "loss": 0.1057, "step": 372 }, { "epoch": 8.674418604651162, "grad_norm": 1.6656956672668457, "learning_rate": 7.093023255813955e-05, "loss": 0.0434, "step": 373 }, { "epoch": 8.69767441860465, "grad_norm": 3.274289608001709, "learning_rate": 7.078488372093023e-05, "loss": 0.1541, "step": 374 }, { "epoch": 8.720930232558139, "grad_norm": 4.472874641418457, "learning_rate": 7.063953488372093e-05, "loss": 0.1322, "step": 375 }, { "epoch": 8.744186046511627, "grad_norm": 2.455209493637085, "learning_rate": 7.049418604651163e-05, "loss": 0.1081, "step": 376 }, { "epoch": 8.767441860465116, "grad_norm": 2.7206978797912598, "learning_rate": 7.034883720930233e-05, "loss": 0.1198, "step": 377 }, { "epoch": 8.790697674418604, "grad_norm": 6.333116054534912, "learning_rate": 7.020348837209303e-05, "loss": 0.1965, "step": 378 }, { "epoch": 8.813953488372093, "grad_norm": 1.0274301767349243, "learning_rate": 7.005813953488373e-05, "loss": 0.0388, "step": 379 }, { "epoch": 8.837209302325581, "grad_norm": 4.29472541809082, "learning_rate": 6.991279069767443e-05, "loss": 0.0699, "step": 380 }, { "epoch": 8.837209302325581, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8978290703580035, "eval_loss": 0.26147955656051636, "eval_precision": 0.9003302050747306, "eval_recall": 0.8969156543465151, "eval_runtime": 0.0748, "eval_samples_per_second": 4079.802, "eval_steps_per_second": 66.882, "step": 380 }, { "epoch": 8.86046511627907, "grad_norm": 3.4688355922698975, "learning_rate": 6.976744186046513e-05, "loss": 0.1465, "step": 381 }, { "epoch": 8.883720930232558, "grad_norm": 4.693603992462158, "learning_rate": 6.962209302325582e-05, "loss": 0.0675, "step": 382 }, { "epoch": 8.906976744186046, "grad_norm": 1.98982572555542, "learning_rate": 6.947674418604651e-05, "loss": 0.1484, "step": 383 }, { "epoch": 8.930232558139535, "grad_norm": 1.312723994255066, "learning_rate": 6.933139534883721e-05, "loss": 0.0458, "step": 384 }, { "epoch": 8.953488372093023, "grad_norm": 2.106917142868042, "learning_rate": 6.918604651162791e-05, "loss": 0.0417, "step": 385 }, { "epoch": 8.976744186046512, "grad_norm": 4.182035446166992, "learning_rate": 6.90406976744186e-05, "loss": 0.0459, "step": 386 }, { "epoch": 9.0, "grad_norm": 2.1285786628723145, "learning_rate": 6.88953488372093e-05, "loss": 0.0743, "step": 387 }, { "epoch": 9.023255813953488, "grad_norm": 4.470812797546387, "learning_rate": 6.875e-05, "loss": 0.0973, "step": 388 }, { "epoch": 9.046511627906977, "grad_norm": 1.6776498556137085, "learning_rate": 6.86046511627907e-05, "loss": 0.1894, "step": 389 }, { "epoch": 9.069767441860465, "grad_norm": 3.9984216690063477, "learning_rate": 6.84593023255814e-05, "loss": 0.1704, "step": 390 }, { "epoch": 9.093023255813954, "grad_norm": 2.3983538150787354, "learning_rate": 6.83139534883721e-05, "loss": 0.0736, "step": 391 }, { "epoch": 9.116279069767442, "grad_norm": 2.989044427871704, "learning_rate": 6.81686046511628e-05, "loss": 0.0892, "step": 392 }, { "epoch": 9.13953488372093, "grad_norm": 3.356992483139038, "learning_rate": 6.802325581395348e-05, "loss": 0.0999, "step": 393 }, { "epoch": 9.162790697674419, "grad_norm": 2.2563319206237793, "learning_rate": 6.787790697674418e-05, "loss": 0.0452, "step": 394 }, { "epoch": 9.186046511627907, "grad_norm": 5.215000152587891, "learning_rate": 6.773255813953488e-05, "loss": 0.0793, "step": 395 }, { "epoch": 9.209302325581396, "grad_norm": 5.427519798278809, "learning_rate": 6.758720930232558e-05, "loss": 0.0755, "step": 396 }, { "epoch": 9.232558139534884, "grad_norm": 7.729520797729492, "learning_rate": 6.744186046511628e-05, "loss": 0.2358, "step": 397 }, { "epoch": 9.255813953488373, "grad_norm": 5.97319221496582, "learning_rate": 6.729651162790698e-05, "loss": 0.1829, "step": 398 }, { "epoch": 9.279069767441861, "grad_norm": 4.662299633026123, "learning_rate": 6.715116279069768e-05, "loss": 0.0549, "step": 399 }, { "epoch": 9.30232558139535, "grad_norm": 5.684605598449707, "learning_rate": 6.700581395348838e-05, "loss": 0.1466, "step": 400 }, { "epoch": 9.30232558139535, "eval_accuracy": 0.9147540983606557, "eval_f1": 0.9145474137931033, "eval_loss": 0.22707484662532806, "eval_precision": 0.9148550724637681, "eval_recall": 0.9143189454639442, "eval_runtime": 0.0754, "eval_samples_per_second": 4044.013, "eval_steps_per_second": 66.295, "step": 400 }, { "epoch": 9.325581395348838, "grad_norm": 8.901346206665039, "learning_rate": 6.686046511627908e-05, "loss": 0.1351, "step": 401 }, { "epoch": 9.348837209302326, "grad_norm": 3.3197011947631836, "learning_rate": 6.671511627906976e-05, "loss": 0.0962, "step": 402 }, { "epoch": 9.372093023255815, "grad_norm": 5.270369529724121, "learning_rate": 6.656976744186046e-05, "loss": 0.0862, "step": 403 }, { "epoch": 9.395348837209303, "grad_norm": 1.4588712453842163, "learning_rate": 6.642441860465116e-05, "loss": 0.0286, "step": 404 }, { "epoch": 9.418604651162791, "grad_norm": 7.928572177886963, "learning_rate": 6.627906976744186e-05, "loss": 0.1137, "step": 405 }, { "epoch": 9.44186046511628, "grad_norm": 4.6285223960876465, "learning_rate": 6.613372093023256e-05, "loss": 0.0907, "step": 406 }, { "epoch": 9.465116279069768, "grad_norm": 1.1047800779342651, "learning_rate": 6.598837209302326e-05, "loss": 0.0279, "step": 407 }, { "epoch": 9.488372093023255, "grad_norm": 2.0423073768615723, "learning_rate": 6.584302325581395e-05, "loss": 0.0528, "step": 408 }, { "epoch": 9.511627906976745, "grad_norm": 3.1448893547058105, "learning_rate": 6.569767441860465e-05, "loss": 0.0557, "step": 409 }, { "epoch": 9.534883720930232, "grad_norm": 3.5887062549591064, "learning_rate": 6.555232558139535e-05, "loss": 0.0858, "step": 410 }, { "epoch": 9.55813953488372, "grad_norm": 4.293898105621338, "learning_rate": 6.540697674418605e-05, "loss": 0.0547, "step": 411 }, { "epoch": 9.581395348837209, "grad_norm": 3.3504300117492676, "learning_rate": 6.526162790697675e-05, "loss": 0.1361, "step": 412 }, { "epoch": 9.604651162790697, "grad_norm": 0.5247292518615723, "learning_rate": 6.511627906976745e-05, "loss": 0.0168, "step": 413 }, { "epoch": 9.627906976744185, "grad_norm": 0.9081286191940308, "learning_rate": 6.497093023255815e-05, "loss": 0.0337, "step": 414 }, { "epoch": 9.651162790697674, "grad_norm": 5.076108932495117, "learning_rate": 6.482558139534885e-05, "loss": 0.0972, "step": 415 }, { "epoch": 9.674418604651162, "grad_norm": 2.2447941303253174, "learning_rate": 6.468023255813955e-05, "loss": 0.0468, "step": 416 }, { "epoch": 9.69767441860465, "grad_norm": 6.019301891326904, "learning_rate": 6.453488372093024e-05, "loss": 0.0967, "step": 417 }, { "epoch": 9.720930232558139, "grad_norm": 1.4316606521606445, "learning_rate": 6.438953488372094e-05, "loss": 0.0666, "step": 418 }, { "epoch": 9.744186046511627, "grad_norm": 1.9364207983016968, "learning_rate": 6.424418604651163e-05, "loss": 0.1047, "step": 419 }, { "epoch": 9.767441860465116, "grad_norm": 4.00551700592041, "learning_rate": 6.409883720930233e-05, "loss": 0.0663, "step": 420 }, { "epoch": 9.767441860465116, "eval_accuracy": 0.9114754098360656, "eval_f1": 0.9106105834464043, "eval_loss": 0.26964467763900757, "eval_precision": 0.9181201550387597, "eval_recall": 0.9089342638063238, "eval_runtime": 0.0726, "eval_samples_per_second": 4201.095, "eval_steps_per_second": 68.87, "step": 420 }, { "epoch": 9.790697674418604, "grad_norm": 3.918790817260742, "learning_rate": 6.395348837209303e-05, "loss": 0.0773, "step": 421 }, { "epoch": 9.813953488372093, "grad_norm": 4.5966901779174805, "learning_rate": 6.380813953488373e-05, "loss": 0.1099, "step": 422 }, { "epoch": 9.837209302325581, "grad_norm": 3.332273244857788, "learning_rate": 6.366279069767442e-05, "loss": 0.0697, "step": 423 }, { "epoch": 9.86046511627907, "grad_norm": 3.300297737121582, "learning_rate": 6.351744186046512e-05, "loss": 0.1235, "step": 424 }, { "epoch": 9.883720930232558, "grad_norm": 1.9253029823303223, "learning_rate": 6.337209302325582e-05, "loss": 0.0276, "step": 425 }, { "epoch": 9.906976744186046, "grad_norm": 2.729339599609375, "learning_rate": 6.322674418604652e-05, "loss": 0.2399, "step": 426 }, { "epoch": 9.930232558139535, "grad_norm": 2.6790566444396973, "learning_rate": 6.308139534883722e-05, "loss": 0.0897, "step": 427 }, { "epoch": 9.953488372093023, "grad_norm": 2.726901054382324, "learning_rate": 6.29360465116279e-05, "loss": 0.0768, "step": 428 }, { "epoch": 9.976744186046512, "grad_norm": 2.4175329208374023, "learning_rate": 6.27906976744186e-05, "loss": 0.0943, "step": 429 }, { "epoch": 10.0, "grad_norm": 4.533559799194336, "learning_rate": 6.26453488372093e-05, "loss": 0.127, "step": 430 }, { "epoch": 10.023255813953488, "grad_norm": 3.8240880966186523, "learning_rate": 6.25e-05, "loss": 0.083, "step": 431 }, { "epoch": 10.046511627906977, "grad_norm": 2.4785892963409424, "learning_rate": 6.23546511627907e-05, "loss": 0.0292, "step": 432 }, { "epoch": 10.069767441860465, "grad_norm": 1.059528112411499, "learning_rate": 6.22093023255814e-05, "loss": 0.0619, "step": 433 }, { "epoch": 10.093023255813954, "grad_norm": 2.8850722312927246, "learning_rate": 6.20639534883721e-05, "loss": 0.1071, "step": 434 }, { "epoch": 10.116279069767442, "grad_norm": 1.9426606893539429, "learning_rate": 6.19186046511628e-05, "loss": 0.0988, "step": 435 }, { "epoch": 10.13953488372093, "grad_norm": 2.1318485736846924, "learning_rate": 6.17732558139535e-05, "loss": 0.1346, "step": 436 }, { "epoch": 10.162790697674419, "grad_norm": 2.793272018432617, "learning_rate": 6.162790697674418e-05, "loss": 0.0296, "step": 437 }, { "epoch": 10.186046511627907, "grad_norm": 1.0038514137268066, "learning_rate": 6.148255813953488e-05, "loss": 0.0576, "step": 438 }, { "epoch": 10.209302325581396, "grad_norm": 2.623504400253296, "learning_rate": 6.133720930232558e-05, "loss": 0.0607, "step": 439 }, { "epoch": 10.232558139534884, "grad_norm": 1.1035139560699463, "learning_rate": 6.119186046511628e-05, "loss": 0.0473, "step": 440 }, { "epoch": 10.232558139534884, "eval_accuracy": 0.9049180327868852, "eval_f1": 0.9045074224021592, "eval_loss": 0.2681082785129547, "eval_precision": 0.9062147872063795, "eval_recall": 0.9037649694150082, "eval_runtime": 0.0732, "eval_samples_per_second": 4164.7, "eval_steps_per_second": 68.274, "step": 440 }, { "epoch": 10.255813953488373, "grad_norm": 4.833049774169922, "learning_rate": 6.104651162790698e-05, "loss": 0.0978, "step": 441 }, { "epoch": 10.279069767441861, "grad_norm": 0.8670517802238464, "learning_rate": 6.0901162790697675e-05, "loss": 0.0284, "step": 442 }, { "epoch": 10.30232558139535, "grad_norm": 4.661880016326904, "learning_rate": 6.0755813953488374e-05, "loss": 0.1335, "step": 443 }, { "epoch": 10.325581395348838, "grad_norm": 4.113884449005127, "learning_rate": 6.0610465116279066e-05, "loss": 0.0624, "step": 444 }, { "epoch": 10.348837209302326, "grad_norm": 4.014869689941406, "learning_rate": 6.0465116279069765e-05, "loss": 0.075, "step": 445 }, { "epoch": 10.372093023255815, "grad_norm": 0.5572329759597778, "learning_rate": 6.0319767441860464e-05, "loss": 0.0172, "step": 446 }, { "epoch": 10.395348837209303, "grad_norm": 2.2173776626586914, "learning_rate": 6.017441860465116e-05, "loss": 0.0753, "step": 447 }, { "epoch": 10.418604651162791, "grad_norm": 4.567299842834473, "learning_rate": 6.002906976744186e-05, "loss": 0.1678, "step": 448 }, { "epoch": 10.44186046511628, "grad_norm": 2.608336925506592, "learning_rate": 5.9883720930232554e-05, "loss": 0.1871, "step": 449 }, { "epoch": 10.465116279069768, "grad_norm": 2.773268699645996, "learning_rate": 5.973837209302325e-05, "loss": 0.1034, "step": 450 }, { "epoch": 10.488372093023255, "grad_norm": 4.1804304122924805, "learning_rate": 5.959302325581395e-05, "loss": 0.1966, "step": 451 }, { "epoch": 10.511627906976745, "grad_norm": 2.119896650314331, "learning_rate": 5.944767441860465e-05, "loss": 0.0443, "step": 452 }, { "epoch": 10.534883720930232, "grad_norm": 2.4072678089141846, "learning_rate": 5.9302325581395356e-05, "loss": 0.0509, "step": 453 }, { "epoch": 10.55813953488372, "grad_norm": 2.580864906311035, "learning_rate": 5.9156976744186055e-05, "loss": 0.1404, "step": 454 }, { "epoch": 10.581395348837209, "grad_norm": 1.998783826828003, "learning_rate": 5.9011627906976754e-05, "loss": 0.0546, "step": 455 }, { "epoch": 10.604651162790697, "grad_norm": 1.9264705181121826, "learning_rate": 5.886627906976745e-05, "loss": 0.0753, "step": 456 }, { "epoch": 10.627906976744185, "grad_norm": 1.8207919597625732, "learning_rate": 5.8720930232558145e-05, "loss": 0.0783, "step": 457 }, { "epoch": 10.651162790697674, "grad_norm": 3.059663772583008, "learning_rate": 5.8575581395348844e-05, "loss": 0.0829, "step": 458 }, { "epoch": 10.674418604651162, "grad_norm": 2.6100552082061768, "learning_rate": 5.843023255813954e-05, "loss": 0.1253, "step": 459 }, { "epoch": 10.69767441860465, "grad_norm": 2.3080873489379883, "learning_rate": 5.828488372093024e-05, "loss": 0.0337, "step": 460 }, { "epoch": 10.69767441860465, "eval_accuracy": 0.8918032786885246, "eval_f1": 0.8916355336398186, "eval_loss": 0.2402210682630539, "eval_precision": 0.8915439593558943, "eval_recall": 0.8917463599551995, "eval_runtime": 0.0746, "eval_samples_per_second": 4086.997, "eval_steps_per_second": 67.0, "step": 460 }, { "epoch": 10.720930232558139, "grad_norm": 1.366176724433899, "learning_rate": 5.8139534883720933e-05, "loss": 0.0355, "step": 461 }, { "epoch": 10.744186046511627, "grad_norm": 3.6105799674987793, "learning_rate": 5.799418604651163e-05, "loss": 0.1579, "step": 462 }, { "epoch": 10.767441860465116, "grad_norm": 2.1533732414245605, "learning_rate": 5.784883720930233e-05, "loss": 0.0605, "step": 463 }, { "epoch": 10.790697674418604, "grad_norm": 2.887242078781128, "learning_rate": 5.770348837209303e-05, "loss": 0.0787, "step": 464 }, { "epoch": 10.813953488372093, "grad_norm": 4.048674583435059, "learning_rate": 5.755813953488373e-05, "loss": 0.0646, "step": 465 }, { "epoch": 10.837209302325581, "grad_norm": 4.225156307220459, "learning_rate": 5.741279069767442e-05, "loss": 0.0449, "step": 466 }, { "epoch": 10.86046511627907, "grad_norm": 5.0084123611450195, "learning_rate": 5.726744186046512e-05, "loss": 0.0798, "step": 467 }, { "epoch": 10.883720930232558, "grad_norm": 2.2451436519622803, "learning_rate": 5.712209302325582e-05, "loss": 0.1, "step": 468 }, { "epoch": 10.906976744186046, "grad_norm": 2.683537006378174, "learning_rate": 5.697674418604652e-05, "loss": 0.0875, "step": 469 }, { "epoch": 10.930232558139535, "grad_norm": 2.50972580909729, "learning_rate": 5.683139534883721e-05, "loss": 0.0313, "step": 470 }, { "epoch": 10.953488372093023, "grad_norm": 8.650345802307129, "learning_rate": 5.668604651162791e-05, "loss": 0.0685, "step": 471 }, { "epoch": 10.976744186046512, "grad_norm": 2.9768123626708984, "learning_rate": 5.654069767441861e-05, "loss": 0.031, "step": 472 }, { "epoch": 11.0, "grad_norm": 2.2081665992736816, "learning_rate": 5.6395348837209306e-05, "loss": 0.0335, "step": 473 }, { "epoch": 11.023255813953488, "grad_norm": 4.979495048522949, "learning_rate": 5.6250000000000005e-05, "loss": 0.0331, "step": 474 }, { "epoch": 11.046511627906977, "grad_norm": 2.000579833984375, "learning_rate": 5.61046511627907e-05, "loss": 0.0472, "step": 475 }, { "epoch": 11.069767441860465, "grad_norm": 2.3896875381469727, "learning_rate": 5.5959302325581396e-05, "loss": 0.0745, "step": 476 }, { "epoch": 11.093023255813954, "grad_norm": 2.0320870876312256, "learning_rate": 5.5813953488372095e-05, "loss": 0.0634, "step": 477 }, { "epoch": 11.116279069767442, "grad_norm": 2.134622812271118, "learning_rate": 5.5668604651162794e-05, "loss": 0.0387, "step": 478 }, { "epoch": 11.13953488372093, "grad_norm": 1.0899347066879272, "learning_rate": 5.552325581395349e-05, "loss": 0.0455, "step": 479 }, { "epoch": 11.162790697674419, "grad_norm": 0.905604898929596, "learning_rate": 5.5377906976744185e-05, "loss": 0.0156, "step": 480 }, { "epoch": 11.162790697674419, "eval_accuracy": 0.9016393442622951, "eval_f1": 0.9008624425760596, "eval_loss": 0.3055068254470825, "eval_precision": 0.906113154668068, "eval_recall": 0.8995003015421728, "eval_runtime": 0.0798, "eval_samples_per_second": 3820.2, "eval_steps_per_second": 62.626, "step": 480 }, { "epoch": 11.186046511627907, "grad_norm": 1.7722190618515015, "learning_rate": 5.5232558139534884e-05, "loss": 0.0672, "step": 481 }, { "epoch": 11.209302325581396, "grad_norm": 3.3152947425842285, "learning_rate": 5.508720930232558e-05, "loss": 0.1093, "step": 482 }, { "epoch": 11.232558139534884, "grad_norm": 3.6544511318206787, "learning_rate": 5.494186046511628e-05, "loss": 0.0894, "step": 483 }, { "epoch": 11.255813953488373, "grad_norm": 4.428676128387451, "learning_rate": 5.4796511627906974e-05, "loss": 0.0495, "step": 484 }, { "epoch": 11.279069767441861, "grad_norm": 3.5235519409179688, "learning_rate": 5.465116279069767e-05, "loss": 0.0313, "step": 485 }, { "epoch": 11.30232558139535, "grad_norm": 1.8089061975479126, "learning_rate": 5.450581395348837e-05, "loss": 0.0294, "step": 486 }, { "epoch": 11.325581395348838, "grad_norm": 9.173127174377441, "learning_rate": 5.436046511627907e-05, "loss": 0.0559, "step": 487 }, { "epoch": 11.348837209302326, "grad_norm": 10.084165573120117, "learning_rate": 5.421511627906977e-05, "loss": 0.0759, "step": 488 }, { "epoch": 11.372093023255815, "grad_norm": 0.5530898571014404, "learning_rate": 5.406976744186046e-05, "loss": 0.0202, "step": 489 }, { "epoch": 11.395348837209303, "grad_norm": 4.215567588806152, "learning_rate": 5.392441860465116e-05, "loss": 0.1209, "step": 490 }, { "epoch": 11.418604651162791, "grad_norm": 0.7583315372467041, "learning_rate": 5.377906976744186e-05, "loss": 0.0054, "step": 491 }, { "epoch": 11.44186046511628, "grad_norm": 0.2435784786939621, "learning_rate": 5.363372093023256e-05, "loss": 0.0064, "step": 492 }, { "epoch": 11.465116279069768, "grad_norm": 2.534301519393921, "learning_rate": 5.348837209302326e-05, "loss": 0.0733, "step": 493 }, { "epoch": 11.488372093023255, "grad_norm": 3.94118332862854, "learning_rate": 5.334302325581395e-05, "loss": 0.0812, "step": 494 }, { "epoch": 11.511627906976745, "grad_norm": 3.799800157546997, "learning_rate": 5.319767441860465e-05, "loss": 0.1377, "step": 495 }, { "epoch": 11.534883720930232, "grad_norm": 0.16743521392345428, "learning_rate": 5.305232558139536e-05, "loss": 0.0054, "step": 496 }, { "epoch": 11.55813953488372, "grad_norm": 1.3862305879592896, "learning_rate": 5.290697674418605e-05, "loss": 0.0364, "step": 497 }, { "epoch": 11.581395348837209, "grad_norm": 3.8730082511901855, "learning_rate": 5.276162790697675e-05, "loss": 0.0827, "step": 498 }, { "epoch": 11.604651162790697, "grad_norm": 5.6961188316345215, "learning_rate": 5.261627906976745e-05, "loss": 0.0936, "step": 499 }, { "epoch": 11.627906976744185, "grad_norm": 0.9591822028160095, "learning_rate": 5.247093023255815e-05, "loss": 0.0146, "step": 500 }, { "epoch": 11.627906976744185, "eval_accuracy": 0.9016393442622951, "eval_f1": 0.900974025974026, "eval_loss": 0.32393914461135864, "eval_precision": 0.9050144016758314, "eval_recall": 0.899780304988369, "eval_runtime": 0.075, "eval_samples_per_second": 4066.651, "eval_steps_per_second": 66.666, "step": 500 }, { "epoch": 11.651162790697674, "grad_norm": 5.92799186706543, "learning_rate": 5.232558139534884e-05, "loss": 0.0904, "step": 501 }, { "epoch": 11.674418604651162, "grad_norm": 9.12846851348877, "learning_rate": 5.218023255813954e-05, "loss": 0.322, "step": 502 }, { "epoch": 11.69767441860465, "grad_norm": 2.95119571685791, "learning_rate": 5.203488372093024e-05, "loss": 0.0175, "step": 503 }, { "epoch": 11.720930232558139, "grad_norm": 3.9121649265289307, "learning_rate": 5.188953488372094e-05, "loss": 0.0733, "step": 504 }, { "epoch": 11.744186046511627, "grad_norm": 2.7404439449310303, "learning_rate": 5.1744186046511636e-05, "loss": 0.0178, "step": 505 }, { "epoch": 11.767441860465116, "grad_norm": 4.207398891448975, "learning_rate": 5.159883720930233e-05, "loss": 0.1179, "step": 506 }, { "epoch": 11.790697674418604, "grad_norm": 5.63287878036499, "learning_rate": 5.145348837209303e-05, "loss": 0.0728, "step": 507 }, { "epoch": 11.813953488372093, "grad_norm": 4.2650556564331055, "learning_rate": 5.1308139534883726e-05, "loss": 0.0409, "step": 508 }, { "epoch": 11.837209302325581, "grad_norm": 4.405393123626709, "learning_rate": 5.1162790697674425e-05, "loss": 0.1272, "step": 509 }, { "epoch": 11.86046511627907, "grad_norm": 3.2369868755340576, "learning_rate": 5.1017441860465124e-05, "loss": 0.1354, "step": 510 }, { "epoch": 11.883720930232558, "grad_norm": 2.492950439453125, "learning_rate": 5.0872093023255816e-05, "loss": 0.116, "step": 511 }, { "epoch": 11.906976744186046, "grad_norm": 5.160562992095947, "learning_rate": 5.0726744186046515e-05, "loss": 0.0813, "step": 512 }, { "epoch": 11.930232558139535, "grad_norm": 3.234938383102417, "learning_rate": 5.0581395348837214e-05, "loss": 0.041, "step": 513 }, { "epoch": 11.953488372093023, "grad_norm": 2.3901174068450928, "learning_rate": 5.043604651162791e-05, "loss": 0.0285, "step": 514 }, { "epoch": 11.976744186046512, "grad_norm": 1.9587090015411377, "learning_rate": 5.0290697674418605e-05, "loss": 0.0191, "step": 515 }, { "epoch": 12.0, "grad_norm": 1.5882123708724976, "learning_rate": 5.0145348837209304e-05, "loss": 0.0292, "step": 516 }, { "epoch": 12.023255813953488, "grad_norm": 1.0756316184997559, "learning_rate": 5e-05, "loss": 0.0348, "step": 517 }, { "epoch": 12.046511627906977, "grad_norm": 7.145483016967773, "learning_rate": 4.98546511627907e-05, "loss": 0.1204, "step": 518 }, { "epoch": 12.069767441860465, "grad_norm": 10.459267616271973, "learning_rate": 4.97093023255814e-05, "loss": 0.0682, "step": 519 }, { "epoch": 12.093023255813954, "grad_norm": 2.6111533641815186, "learning_rate": 4.956395348837209e-05, "loss": 0.1088, "step": 520 }, { "epoch": 12.093023255813954, "eval_accuracy": 0.9081967213114754, "eval_f1": 0.907471613070989, "eval_loss": 0.29710283875465393, "eval_precision": 0.9127912068663514, "eval_recall": 0.9060696131644697, "eval_runtime": 0.0726, "eval_samples_per_second": 4200.336, "eval_steps_per_second": 68.858, "step": 520 }, { "epoch": 12.116279069767442, "grad_norm": 1.2339119911193848, "learning_rate": 4.941860465116279e-05, "loss": 0.0222, "step": 521 }, { "epoch": 12.13953488372093, "grad_norm": 1.6148200035095215, "learning_rate": 4.927325581395349e-05, "loss": 0.0968, "step": 522 }, { "epoch": 12.162790697674419, "grad_norm": 2.2320408821105957, "learning_rate": 4.912790697674419e-05, "loss": 0.0742, "step": 523 }, { "epoch": 12.186046511627907, "grad_norm": 1.0352402925491333, "learning_rate": 4.898255813953488e-05, "loss": 0.0287, "step": 524 }, { "epoch": 12.209302325581396, "grad_norm": 3.2948336601257324, "learning_rate": 4.883720930232558e-05, "loss": 0.0945, "step": 525 }, { "epoch": 12.232558139534884, "grad_norm": 0.48531392216682434, "learning_rate": 4.869186046511628e-05, "loss": 0.0119, "step": 526 }, { "epoch": 12.255813953488373, "grad_norm": 3.030161142349243, "learning_rate": 4.854651162790698e-05, "loss": 0.0713, "step": 527 }, { "epoch": 12.279069767441861, "grad_norm": 1.6946715116500854, "learning_rate": 4.8401162790697676e-05, "loss": 0.0322, "step": 528 }, { "epoch": 12.30232558139535, "grad_norm": 4.674280643463135, "learning_rate": 4.8255813953488375e-05, "loss": 0.048, "step": 529 }, { "epoch": 12.325581395348838, "grad_norm": 1.2208576202392578, "learning_rate": 4.8110465116279074e-05, "loss": 0.009, "step": 530 }, { "epoch": 12.348837209302326, "grad_norm": 7.181972026824951, "learning_rate": 4.796511627906977e-05, "loss": 0.0275, "step": 531 }, { "epoch": 12.372093023255815, "grad_norm": 1.5175693035125732, "learning_rate": 4.781976744186047e-05, "loss": 0.0246, "step": 532 }, { "epoch": 12.395348837209303, "grad_norm": 5.098257064819336, "learning_rate": 4.7674418604651164e-05, "loss": 0.0758, "step": 533 }, { "epoch": 12.418604651162791, "grad_norm": 2.7552084922790527, "learning_rate": 4.752906976744186e-05, "loss": 0.0558, "step": 534 }, { "epoch": 12.44186046511628, "grad_norm": 2.456735134124756, "learning_rate": 4.738372093023256e-05, "loss": 0.0484, "step": 535 }, { "epoch": 12.465116279069768, "grad_norm": 1.6237318515777588, "learning_rate": 4.723837209302326e-05, "loss": 0.08, "step": 536 }, { "epoch": 12.488372093023255, "grad_norm": 3.8842220306396484, "learning_rate": 4.709302325581396e-05, "loss": 0.0858, "step": 537 }, { "epoch": 12.511627906976745, "grad_norm": 3.0373013019561768, "learning_rate": 4.694767441860465e-05, "loss": 0.0256, "step": 538 }, { "epoch": 12.534883720930232, "grad_norm": 5.356001853942871, "learning_rate": 4.680232558139535e-05, "loss": 0.1057, "step": 539 }, { "epoch": 12.55813953488372, "grad_norm": 0.7480242252349854, "learning_rate": 4.665697674418605e-05, "loss": 0.0102, "step": 540 }, { "epoch": 12.55813953488372, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8983213068212369, "eval_loss": 0.27956801652908325, "eval_precision": 0.8984348125214998, "eval_recall": 0.8991556819160851, "eval_runtime": 0.0733, "eval_samples_per_second": 4158.554, "eval_steps_per_second": 68.173, "step": 540 }, { "epoch": 12.581395348837209, "grad_norm": 8.634737014770508, "learning_rate": 4.651162790697675e-05, "loss": 0.1041, "step": 541 }, { "epoch": 12.604651162790697, "grad_norm": 2.154197931289673, "learning_rate": 4.636627906976744e-05, "loss": 0.0883, "step": 542 }, { "epoch": 12.627906976744185, "grad_norm": 5.6526265144348145, "learning_rate": 4.622093023255814e-05, "loss": 0.0634, "step": 543 }, { "epoch": 12.651162790697674, "grad_norm": 0.4163094162940979, "learning_rate": 4.607558139534884e-05, "loss": 0.0102, "step": 544 }, { "epoch": 12.674418604651162, "grad_norm": 2.6310293674468994, "learning_rate": 4.593023255813954e-05, "loss": 0.0613, "step": 545 }, { "epoch": 12.69767441860465, "grad_norm": 0.8978974223136902, "learning_rate": 4.5784883720930236e-05, "loss": 0.0118, "step": 546 }, { "epoch": 12.720930232558139, "grad_norm": 3.0897459983825684, "learning_rate": 4.563953488372093e-05, "loss": 0.0592, "step": 547 }, { "epoch": 12.744186046511627, "grad_norm": 1.5553648471832275, "learning_rate": 4.549418604651163e-05, "loss": 0.0357, "step": 548 }, { "epoch": 12.767441860465116, "grad_norm": 1.9957389831542969, "learning_rate": 4.5348837209302326e-05, "loss": 0.0415, "step": 549 }, { "epoch": 12.790697674418604, "grad_norm": 0.22972577810287476, "learning_rate": 4.520348837209303e-05, "loss": 0.0078, "step": 550 }, { "epoch": 12.813953488372093, "grad_norm": 2.78206205368042, "learning_rate": 4.505813953488372e-05, "loss": 0.026, "step": 551 }, { "epoch": 12.837209302325581, "grad_norm": 3.2643375396728516, "learning_rate": 4.491279069767442e-05, "loss": 0.0386, "step": 552 }, { "epoch": 12.86046511627907, "grad_norm": 9.664105415344238, "learning_rate": 4.476744186046512e-05, "loss": 0.1431, "step": 553 }, { "epoch": 12.883720930232558, "grad_norm": 3.2015540599823, "learning_rate": 4.462209302325582e-05, "loss": 0.0605, "step": 554 }, { "epoch": 12.906976744186046, "grad_norm": 4.17910099029541, "learning_rate": 4.447674418604651e-05, "loss": 0.0797, "step": 555 }, { "epoch": 12.930232558139535, "grad_norm": 3.999985933303833, "learning_rate": 4.433139534883721e-05, "loss": 0.117, "step": 556 }, { "epoch": 12.953488372093023, "grad_norm": 1.1732484102249146, "learning_rate": 4.418604651162791e-05, "loss": 0.0244, "step": 557 }, { "epoch": 12.976744186046512, "grad_norm": 4.184539794921875, "learning_rate": 4.404069767441861e-05, "loss": 0.0517, "step": 558 }, { "epoch": 13.0, "grad_norm": 2.3533170223236084, "learning_rate": 4.389534883720931e-05, "loss": 0.0505, "step": 559 }, { "epoch": 13.023255813953488, "grad_norm": 2.3821420669555664, "learning_rate": 4.375e-05, "loss": 0.0253, "step": 560 }, { "epoch": 13.023255813953488, "eval_accuracy": 0.8885245901639345, "eval_f1": 0.8879936058066189, "eval_loss": 0.3079434633255005, "eval_precision": 0.8900242992276317, "eval_recall": 0.8872016886361678, "eval_runtime": 0.0743, "eval_samples_per_second": 4102.318, "eval_steps_per_second": 67.251, "step": 560 }, { "epoch": 13.046511627906977, "grad_norm": 3.6151607036590576, "learning_rate": 4.36046511627907e-05, "loss": 0.08, "step": 561 }, { "epoch": 13.069767441860465, "grad_norm": 1.870180368423462, "learning_rate": 4.34593023255814e-05, "loss": 0.0475, "step": 562 }, { "epoch": 13.093023255813954, "grad_norm": 3.9536831378936768, "learning_rate": 4.3313953488372096e-05, "loss": 0.0206, "step": 563 }, { "epoch": 13.116279069767442, "grad_norm": 6.368700981140137, "learning_rate": 4.3168604651162795e-05, "loss": 0.0485, "step": 564 }, { "epoch": 13.13953488372093, "grad_norm": 1.1507411003112793, "learning_rate": 4.302325581395349e-05, "loss": 0.0133, "step": 565 }, { "epoch": 13.162790697674419, "grad_norm": 1.034654140472412, "learning_rate": 4.2877906976744186e-05, "loss": 0.0437, "step": 566 }, { "epoch": 13.186046511627907, "grad_norm": 0.2083161622285843, "learning_rate": 4.2732558139534885e-05, "loss": 0.0054, "step": 567 }, { "epoch": 13.209302325581396, "grad_norm": 1.1866053342819214, "learning_rate": 4.2587209302325584e-05, "loss": 0.0134, "step": 568 }, { "epoch": 13.232558139534884, "grad_norm": 0.6963038444519043, "learning_rate": 4.2441860465116276e-05, "loss": 0.0083, "step": 569 }, { "epoch": 13.255813953488373, "grad_norm": 3.132704973220825, "learning_rate": 4.2296511627906975e-05, "loss": 0.0567, "step": 570 }, { "epoch": 13.279069767441861, "grad_norm": 1.62773859500885, "learning_rate": 4.215116279069768e-05, "loss": 0.0803, "step": 571 }, { "epoch": 13.30232558139535, "grad_norm": 0.7541981935501099, "learning_rate": 4.200581395348838e-05, "loss": 0.0114, "step": 572 }, { "epoch": 13.325581395348838, "grad_norm": 5.177672386169434, "learning_rate": 4.186046511627907e-05, "loss": 0.0302, "step": 573 }, { "epoch": 13.348837209302326, "grad_norm": 3.653960704803467, "learning_rate": 4.171511627906977e-05, "loss": 0.0648, "step": 574 }, { "epoch": 13.372093023255815, "grad_norm": 0.6348634362220764, "learning_rate": 4.156976744186047e-05, "loss": 0.0077, "step": 575 }, { "epoch": 13.395348837209303, "grad_norm": 4.980737686157227, "learning_rate": 4.142441860465117e-05, "loss": 0.0857, "step": 576 }, { "epoch": 13.418604651162791, "grad_norm": 4.5483551025390625, "learning_rate": 4.127906976744187e-05, "loss": 0.0456, "step": 577 }, { "epoch": 13.44186046511628, "grad_norm": 4.054844379425049, "learning_rate": 4.113372093023256e-05, "loss": 0.0837, "step": 578 }, { "epoch": 13.465116279069768, "grad_norm": 3.3397719860076904, "learning_rate": 4.098837209302326e-05, "loss": 0.0742, "step": 579 }, { "epoch": 13.488372093023255, "grad_norm": 6.787147045135498, "learning_rate": 4.0843023255813957e-05, "loss": 0.0868, "step": 580 }, { "epoch": 13.488372093023255, "eval_accuracy": 0.8918032786885246, "eval_f1": 0.8914251194701244, "eval_loss": 0.34987983107566833, "eval_precision": 0.8924277806607853, "eval_recall": 0.8909063496166107, "eval_runtime": 0.0747, "eval_samples_per_second": 4085.209, "eval_steps_per_second": 66.971, "step": 580 }, { "epoch": 13.511627906976745, "grad_norm": 1.7752641439437866, "learning_rate": 4.0697674418604655e-05, "loss": 0.0476, "step": 581 }, { "epoch": 13.534883720930232, "grad_norm": 2.80531907081604, "learning_rate": 4.055232558139535e-05, "loss": 0.0202, "step": 582 }, { "epoch": 13.55813953488372, "grad_norm": 0.5729110836982727, "learning_rate": 4.0406976744186046e-05, "loss": 0.0078, "step": 583 }, { "epoch": 13.581395348837209, "grad_norm": 5.221317768096924, "learning_rate": 4.0261627906976745e-05, "loss": 0.092, "step": 584 }, { "epoch": 13.604651162790697, "grad_norm": 3.757556915283203, "learning_rate": 4.0116279069767444e-05, "loss": 0.024, "step": 585 }, { "epoch": 13.627906976744185, "grad_norm": 6.441154479980469, "learning_rate": 3.997093023255814e-05, "loss": 0.1422, "step": 586 }, { "epoch": 13.651162790697674, "grad_norm": 2.314418077468872, "learning_rate": 3.9825581395348835e-05, "loss": 0.0815, "step": 587 }, { "epoch": 13.674418604651162, "grad_norm": 2.129340887069702, "learning_rate": 3.9680232558139534e-05, "loss": 0.0419, "step": 588 }, { "epoch": 13.69767441860465, "grad_norm": 1.477169394493103, "learning_rate": 3.953488372093023e-05, "loss": 0.0379, "step": 589 }, { "epoch": 13.720930232558139, "grad_norm": 2.567399740219116, "learning_rate": 3.938953488372093e-05, "loss": 0.1169, "step": 590 }, { "epoch": 13.744186046511627, "grad_norm": 1.2707808017730713, "learning_rate": 3.924418604651163e-05, "loss": 0.0225, "step": 591 }, { "epoch": 13.767441860465116, "grad_norm": 0.3642787039279938, "learning_rate": 3.909883720930232e-05, "loss": 0.0053, "step": 592 }, { "epoch": 13.790697674418604, "grad_norm": 2.2786664962768555, "learning_rate": 3.895348837209303e-05, "loss": 0.043, "step": 593 }, { "epoch": 13.813953488372093, "grad_norm": 6.119050979614258, "learning_rate": 3.880813953488373e-05, "loss": 0.1129, "step": 594 }, { "epoch": 13.837209302325581, "grad_norm": 3.127246141433716, "learning_rate": 3.866279069767442e-05, "loss": 0.1094, "step": 595 }, { "epoch": 13.86046511627907, "grad_norm": 4.50978946685791, "learning_rate": 3.851744186046512e-05, "loss": 0.0714, "step": 596 }, { "epoch": 13.883720930232558, "grad_norm": 2.3526670932769775, "learning_rate": 3.837209302325582e-05, "loss": 0.1126, "step": 597 }, { "epoch": 13.906976744186046, "grad_norm": 4.066991806030273, "learning_rate": 3.8226744186046516e-05, "loss": 0.0466, "step": 598 }, { "epoch": 13.930232558139535, "grad_norm": 1.5120619535446167, "learning_rate": 3.8081395348837215e-05, "loss": 0.0609, "step": 599 }, { "epoch": 13.953488372093023, "grad_norm": 3.2349016666412354, "learning_rate": 3.793604651162791e-05, "loss": 0.0399, "step": 600 }, { "epoch": 13.953488372093023, "eval_accuracy": 0.8885245901639345, "eval_f1": 0.8880903047569713, "eval_loss": 0.36928993463516235, "eval_precision": 0.8893939393939394, "eval_recall": 0.887481692082364, "eval_runtime": 0.0753, "eval_samples_per_second": 4048.646, "eval_steps_per_second": 66.371, "step": 600 }, { "epoch": 13.976744186046512, "grad_norm": 2.7635951042175293, "learning_rate": 3.7790697674418606e-05, "loss": 0.1149, "step": 601 }, { "epoch": 14.0, "grad_norm": 0.716441810131073, "learning_rate": 3.7645348837209305e-05, "loss": 0.0095, "step": 602 }, { "epoch": 14.023255813953488, "grad_norm": 1.099993348121643, "learning_rate": 3.7500000000000003e-05, "loss": 0.0179, "step": 603 }, { "epoch": 14.046511627906977, "grad_norm": 0.9968247413635254, "learning_rate": 3.73546511627907e-05, "loss": 0.0295, "step": 604 }, { "epoch": 14.069767441860465, "grad_norm": 0.9443352222442627, "learning_rate": 3.7209302325581394e-05, "loss": 0.0121, "step": 605 }, { "epoch": 14.093023255813954, "grad_norm": 0.24217644333839417, "learning_rate": 3.706395348837209e-05, "loss": 0.0066, "step": 606 }, { "epoch": 14.116279069767442, "grad_norm": 0.32041826844215393, "learning_rate": 3.691860465116279e-05, "loss": 0.0071, "step": 607 }, { "epoch": 14.13953488372093, "grad_norm": 4.715015411376953, "learning_rate": 3.677325581395349e-05, "loss": 0.0775, "step": 608 }, { "epoch": 14.162790697674419, "grad_norm": 1.9597333669662476, "learning_rate": 3.662790697674418e-05, "loss": 0.0232, "step": 609 }, { "epoch": 14.186046511627907, "grad_norm": 1.3680325746536255, "learning_rate": 3.648255813953488e-05, "loss": 0.0194, "step": 610 }, { "epoch": 14.209302325581396, "grad_norm": 7.5977396965026855, "learning_rate": 3.633720930232558e-05, "loss": 0.1274, "step": 611 }, { "epoch": 14.232558139534884, "grad_norm": 7.637692451477051, "learning_rate": 3.619186046511628e-05, "loss": 0.1341, "step": 612 }, { "epoch": 14.255813953488373, "grad_norm": 10.332743644714355, "learning_rate": 3.604651162790698e-05, "loss": 0.1039, "step": 613 }, { "epoch": 14.279069767441861, "grad_norm": 1.6394683122634888, "learning_rate": 3.590116279069768e-05, "loss": 0.0126, "step": 614 }, { "epoch": 14.30232558139535, "grad_norm": 1.69204580783844, "learning_rate": 3.5755813953488376e-05, "loss": 0.0327, "step": 615 }, { "epoch": 14.325581395348838, "grad_norm": 3.1536524295806885, "learning_rate": 3.5610465116279075e-05, "loss": 0.0347, "step": 616 }, { "epoch": 14.348837209302326, "grad_norm": 0.6318650841712952, "learning_rate": 3.5465116279069774e-05, "loss": 0.0424, "step": 617 }, { "epoch": 14.372093023255815, "grad_norm": 5.714948654174805, "learning_rate": 3.5319767441860466e-05, "loss": 0.0339, "step": 618 }, { "epoch": 14.395348837209303, "grad_norm": 4.141314506530762, "learning_rate": 3.5174418604651165e-05, "loss": 0.045, "step": 619 }, { "epoch": 14.418604651162791, "grad_norm": 1.1429768800735474, "learning_rate": 3.5029069767441864e-05, "loss": 0.0291, "step": 620 }, { "epoch": 14.418604651162791, "eval_accuracy": 0.8950819672131147, "eval_f1": 0.8944819235426397, "eval_loss": 0.3720164895057678, "eval_precision": 0.8974286460146188, "eval_recall": 0.8934909968122684, "eval_runtime": 0.0902, "eval_samples_per_second": 3381.6, "eval_steps_per_second": 55.436, "step": 620 }, { "epoch": 14.44186046511628, "grad_norm": 11.266057014465332, "learning_rate": 3.488372093023256e-05, "loss": 0.0526, "step": 621 }, { "epoch": 14.465116279069768, "grad_norm": 0.8715966939926147, "learning_rate": 3.4738372093023255e-05, "loss": 0.0256, "step": 622 }, { "epoch": 14.488372093023255, "grad_norm": 3.1204757690429688, "learning_rate": 3.4593023255813954e-05, "loss": 0.0256, "step": 623 }, { "epoch": 14.511627906976745, "grad_norm": 2.438506841659546, "learning_rate": 3.444767441860465e-05, "loss": 0.0321, "step": 624 }, { "epoch": 14.534883720930232, "grad_norm": 3.6766304969787598, "learning_rate": 3.430232558139535e-05, "loss": 0.1027, "step": 625 }, { "epoch": 14.55813953488372, "grad_norm": 6.79380464553833, "learning_rate": 3.415697674418605e-05, "loss": 0.0597, "step": 626 }, { "epoch": 14.581395348837209, "grad_norm": 1.6378151178359985, "learning_rate": 3.401162790697674e-05, "loss": 0.0238, "step": 627 }, { "epoch": 14.604651162790697, "grad_norm": 0.5451012253761292, "learning_rate": 3.386627906976744e-05, "loss": 0.0103, "step": 628 }, { "epoch": 14.627906976744185, "grad_norm": 2.0606513023376465, "learning_rate": 3.372093023255814e-05, "loss": 0.0223, "step": 629 }, { "epoch": 14.651162790697674, "grad_norm": 0.4602338373661041, "learning_rate": 3.357558139534884e-05, "loss": 0.0057, "step": 630 }, { "epoch": 14.674418604651162, "grad_norm": 4.629768371582031, "learning_rate": 3.343023255813954e-05, "loss": 0.1183, "step": 631 }, { "epoch": 14.69767441860465, "grad_norm": 6.286869525909424, "learning_rate": 3.328488372093023e-05, "loss": 0.0308, "step": 632 }, { "epoch": 14.720930232558139, "grad_norm": 1.118071436882019, "learning_rate": 3.313953488372093e-05, "loss": 0.0692, "step": 633 }, { "epoch": 14.744186046511627, "grad_norm": 2.621478319168091, "learning_rate": 3.299418604651163e-05, "loss": 0.0206, "step": 634 }, { "epoch": 14.767441860465116, "grad_norm": 2.408926486968994, "learning_rate": 3.284883720930233e-05, "loss": 0.0868, "step": 635 }, { "epoch": 14.790697674418604, "grad_norm": 4.2226104736328125, "learning_rate": 3.2703488372093026e-05, "loss": 0.0165, "step": 636 }, { "epoch": 14.813953488372093, "grad_norm": 8.624298095703125, "learning_rate": 3.2558139534883724e-05, "loss": 0.1733, "step": 637 }, { "epoch": 14.837209302325581, "grad_norm": 5.992215633392334, "learning_rate": 3.241279069767442e-05, "loss": 0.0286, "step": 638 }, { "epoch": 14.86046511627907, "grad_norm": 0.8743571043014526, "learning_rate": 3.226744186046512e-05, "loss": 0.0113, "step": 639 }, { "epoch": 14.883720930232558, "grad_norm": 0.9226441979408264, "learning_rate": 3.2122093023255814e-05, "loss": 0.0069, "step": 640 }, { "epoch": 14.883720930232558, "eval_accuracy": 0.8918032786885246, "eval_f1": 0.8915047052355851, "eval_loss": 0.3578951358795166, "eval_precision": 0.8919968919968919, "eval_recall": 0.8911863530628069, "eval_runtime": 0.0756, "eval_samples_per_second": 4035.886, "eval_steps_per_second": 66.162, "step": 640 }, { "epoch": 14.906976744186046, "grad_norm": 4.145755767822266, "learning_rate": 3.197674418604651e-05, "loss": 0.0439, "step": 641 }, { "epoch": 14.930232558139535, "grad_norm": 1.2993311882019043, "learning_rate": 3.183139534883721e-05, "loss": 0.0446, "step": 642 }, { "epoch": 14.953488372093023, "grad_norm": 0.2827296555042267, "learning_rate": 3.168604651162791e-05, "loss": 0.0052, "step": 643 }, { "epoch": 14.976744186046512, "grad_norm": 2.9335291385650635, "learning_rate": 3.154069767441861e-05, "loss": 0.0644, "step": 644 }, { "epoch": 15.0, "grad_norm": 7.6974968910217285, "learning_rate": 3.13953488372093e-05, "loss": 0.0456, "step": 645 }, { "epoch": 15.023255813953488, "grad_norm": 2.137190103530884, "learning_rate": 3.125e-05, "loss": 0.0147, "step": 646 }, { "epoch": 15.046511627906977, "grad_norm": 0.834999680519104, "learning_rate": 3.11046511627907e-05, "loss": 0.0181, "step": 647 }, { "epoch": 15.069767441860465, "grad_norm": 0.8868411183357239, "learning_rate": 3.09593023255814e-05, "loss": 0.0467, "step": 648 }, { "epoch": 15.093023255813954, "grad_norm": 3.1377875804901123, "learning_rate": 3.081395348837209e-05, "loss": 0.0281, "step": 649 }, { "epoch": 15.116279069767442, "grad_norm": 3.3987367153167725, "learning_rate": 3.066860465116279e-05, "loss": 0.1133, "step": 650 }, { "epoch": 15.13953488372093, "grad_norm": 0.19906505942344666, "learning_rate": 3.052325581395349e-05, "loss": 0.0032, "step": 651 }, { "epoch": 15.162790697674419, "grad_norm": 4.011625289916992, "learning_rate": 3.0377906976744187e-05, "loss": 0.0792, "step": 652 }, { "epoch": 15.186046511627907, "grad_norm": 9.421031951904297, "learning_rate": 3.0232558139534883e-05, "loss": 0.0612, "step": 653 }, { "epoch": 15.209302325581396, "grad_norm": 3.5618064403533936, "learning_rate": 3.008720930232558e-05, "loss": 0.0159, "step": 654 }, { "epoch": 15.232558139534884, "grad_norm": 7.569765090942383, "learning_rate": 2.9941860465116277e-05, "loss": 0.1039, "step": 655 }, { "epoch": 15.255813953488373, "grad_norm": 6.453846454620361, "learning_rate": 2.9796511627906976e-05, "loss": 0.0283, "step": 656 }, { "epoch": 15.279069767441861, "grad_norm": 0.1914948970079422, "learning_rate": 2.9651162790697678e-05, "loss": 0.0033, "step": 657 }, { "epoch": 15.30232558139535, "grad_norm": 2.6470947265625, "learning_rate": 2.9505813953488377e-05, "loss": 0.0505, "step": 658 }, { "epoch": 15.325581395348838, "grad_norm": 5.789623737335205, "learning_rate": 2.9360465116279072e-05, "loss": 0.0419, "step": 659 }, { "epoch": 15.348837209302326, "grad_norm": 1.2438925504684448, "learning_rate": 2.921511627906977e-05, "loss": 0.027, "step": 660 }, { "epoch": 15.348837209302326, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8980801776455497, "eval_loss": 0.36847957968711853, "eval_precision": 0.8985798152464819, "eval_recall": 0.8977556646851038, "eval_runtime": 0.0739, "eval_samples_per_second": 4125.669, "eval_steps_per_second": 67.634, "step": 660 }, { "epoch": 15.372093023255815, "grad_norm": 2.8159120082855225, "learning_rate": 2.9069767441860467e-05, "loss": 0.0665, "step": 661 }, { "epoch": 15.395348837209303, "grad_norm": 3.529680013656616, "learning_rate": 2.8924418604651166e-05, "loss": 0.0238, "step": 662 }, { "epoch": 15.418604651162791, "grad_norm": 0.9973799586296082, "learning_rate": 2.8779069767441864e-05, "loss": 0.0051, "step": 663 }, { "epoch": 15.44186046511628, "grad_norm": 2.763955593109131, "learning_rate": 2.863372093023256e-05, "loss": 0.0107, "step": 664 }, { "epoch": 15.465116279069768, "grad_norm": 3.6373209953308105, "learning_rate": 2.848837209302326e-05, "loss": 0.0466, "step": 665 }, { "epoch": 15.488372093023255, "grad_norm": 3.262023448944092, "learning_rate": 2.8343023255813954e-05, "loss": 0.025, "step": 666 }, { "epoch": 15.511627906976745, "grad_norm": 0.7062492370605469, "learning_rate": 2.8197674418604653e-05, "loss": 0.0067, "step": 667 }, { "epoch": 15.534883720930232, "grad_norm": 2.9254820346832275, "learning_rate": 2.805232558139535e-05, "loss": 0.0394, "step": 668 }, { "epoch": 15.55813953488372, "grad_norm": 1.5290218591690063, "learning_rate": 2.7906976744186048e-05, "loss": 0.0268, "step": 669 }, { "epoch": 15.581395348837209, "grad_norm": 3.9581668376922607, "learning_rate": 2.7761627906976746e-05, "loss": 0.0683, "step": 670 }, { "epoch": 15.604651162790697, "grad_norm": 3.121964454650879, "learning_rate": 2.7616279069767442e-05, "loss": 0.0307, "step": 671 }, { "epoch": 15.627906976744185, "grad_norm": 4.223913669586182, "learning_rate": 2.747093023255814e-05, "loss": 0.0778, "step": 672 }, { "epoch": 15.651162790697674, "grad_norm": 1.8996158838272095, "learning_rate": 2.7325581395348836e-05, "loss": 0.0149, "step": 673 }, { "epoch": 15.674418604651162, "grad_norm": 2.5367538928985596, "learning_rate": 2.7180232558139535e-05, "loss": 0.0238, "step": 674 }, { "epoch": 15.69767441860465, "grad_norm": 1.5632944107055664, "learning_rate": 2.703488372093023e-05, "loss": 0.0159, "step": 675 }, { "epoch": 15.720930232558139, "grad_norm": 2.031212329864502, "learning_rate": 2.688953488372093e-05, "loss": 0.0265, "step": 676 }, { "epoch": 15.744186046511627, "grad_norm": 1.0653328895568848, "learning_rate": 2.674418604651163e-05, "loss": 0.0146, "step": 677 }, { "epoch": 15.767441860465116, "grad_norm": 2.078573226928711, "learning_rate": 2.6598837209302324e-05, "loss": 0.0314, "step": 678 }, { "epoch": 15.790697674418604, "grad_norm": 1.004683017730713, "learning_rate": 2.6453488372093026e-05, "loss": 0.0079, "step": 679 }, { "epoch": 15.813953488372093, "grad_norm": 0.6423361897468567, "learning_rate": 2.6308139534883725e-05, "loss": 0.0265, "step": 680 }, { "epoch": 15.813953488372093, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8981460534962133, "eval_loss": 0.3592439889907837, "eval_precision": 0.8982758620689655, "eval_recall": 0.8980356681313001, "eval_runtime": 0.0743, "eval_samples_per_second": 4106.888, "eval_steps_per_second": 67.326, "step": 680 }, { "epoch": 15.837209302325581, "grad_norm": 2.9641597270965576, "learning_rate": 2.616279069767442e-05, "loss": 0.0753, "step": 681 }, { "epoch": 15.86046511627907, "grad_norm": 5.338639259338379, "learning_rate": 2.601744186046512e-05, "loss": 0.0325, "step": 682 }, { "epoch": 15.883720930232558, "grad_norm": 1.9736541509628296, "learning_rate": 2.5872093023255818e-05, "loss": 0.0236, "step": 683 }, { "epoch": 15.906976744186046, "grad_norm": 0.5987362265586853, "learning_rate": 2.5726744186046514e-05, "loss": 0.0056, "step": 684 }, { "epoch": 15.930232558139535, "grad_norm": 1.9034494161605835, "learning_rate": 2.5581395348837212e-05, "loss": 0.095, "step": 685 }, { "epoch": 15.953488372093023, "grad_norm": 0.7597903609275818, "learning_rate": 2.5436046511627908e-05, "loss": 0.0111, "step": 686 }, { "epoch": 15.976744186046512, "grad_norm": 2.3135693073272705, "learning_rate": 2.5290697674418607e-05, "loss": 0.0792, "step": 687 }, { "epoch": 16.0, "grad_norm": 0.07970131933689117, "learning_rate": 2.5145348837209302e-05, "loss": 0.0024, "step": 688 }, { "epoch": 16.023255813953487, "grad_norm": 1.8181146383285522, "learning_rate": 2.5e-05, "loss": 0.0181, "step": 689 }, { "epoch": 16.046511627906977, "grad_norm": 4.323759078979492, "learning_rate": 2.48546511627907e-05, "loss": 0.0386, "step": 690 }, { "epoch": 16.069767441860463, "grad_norm": 3.518137216567993, "learning_rate": 2.4709302325581396e-05, "loss": 0.0794, "step": 691 }, { "epoch": 16.093023255813954, "grad_norm": 0.384950190782547, "learning_rate": 2.4563953488372094e-05, "loss": 0.0042, "step": 692 }, { "epoch": 16.11627906976744, "grad_norm": 0.4682580232620239, "learning_rate": 2.441860465116279e-05, "loss": 0.0078, "step": 693 }, { "epoch": 16.13953488372093, "grad_norm": 1.6920162439346313, "learning_rate": 2.427325581395349e-05, "loss": 0.0181, "step": 694 }, { "epoch": 16.162790697674417, "grad_norm": 3.120163679122925, "learning_rate": 2.4127906976744188e-05, "loss": 0.0414, "step": 695 }, { "epoch": 16.186046511627907, "grad_norm": 0.8569607734680176, "learning_rate": 2.3982558139534887e-05, "loss": 0.0063, "step": 696 }, { "epoch": 16.209302325581394, "grad_norm": 1.5792855024337769, "learning_rate": 2.3837209302325582e-05, "loss": 0.0233, "step": 697 }, { "epoch": 16.232558139534884, "grad_norm": 11.798049926757812, "learning_rate": 2.369186046511628e-05, "loss": 0.031, "step": 698 }, { "epoch": 16.25581395348837, "grad_norm": 2.0980348587036133, "learning_rate": 2.354651162790698e-05, "loss": 0.0345, "step": 699 }, { "epoch": 16.27906976744186, "grad_norm": 1.9684696197509766, "learning_rate": 2.3401162790697675e-05, "loss": 0.0109, "step": 700 }, { "epoch": 16.27906976744186, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8979217273954117, "eval_loss": 0.36475130915641785, "eval_precision": 0.8996056167114501, "eval_recall": 0.8971956577927114, "eval_runtime": 0.0743, "eval_samples_per_second": 4105.148, "eval_steps_per_second": 67.298, "step": 700 }, { "epoch": 16.302325581395348, "grad_norm": 1.2527034282684326, "learning_rate": 2.3255813953488374e-05, "loss": 0.0088, "step": 701 }, { "epoch": 16.325581395348838, "grad_norm": 4.517597198486328, "learning_rate": 2.311046511627907e-05, "loss": 0.05, "step": 702 }, { "epoch": 16.348837209302324, "grad_norm": 2.9931998252868652, "learning_rate": 2.296511627906977e-05, "loss": 0.0852, "step": 703 }, { "epoch": 16.372093023255815, "grad_norm": 1.5361837148666382, "learning_rate": 2.2819767441860464e-05, "loss": 0.0205, "step": 704 }, { "epoch": 16.3953488372093, "grad_norm": 12.71114730834961, "learning_rate": 2.2674418604651163e-05, "loss": 0.0643, "step": 705 }, { "epoch": 16.41860465116279, "grad_norm": 0.1732826828956604, "learning_rate": 2.252906976744186e-05, "loss": 0.0028, "step": 706 }, { "epoch": 16.441860465116278, "grad_norm": 5.568197250366211, "learning_rate": 2.238372093023256e-05, "loss": 0.0147, "step": 707 }, { "epoch": 16.46511627906977, "grad_norm": 3.462963581085205, "learning_rate": 2.2238372093023256e-05, "loss": 0.0639, "step": 708 }, { "epoch": 16.488372093023255, "grad_norm": 0.7707027792930603, "learning_rate": 2.2093023255813955e-05, "loss": 0.0052, "step": 709 }, { "epoch": 16.511627906976745, "grad_norm": 2.6119136810302734, "learning_rate": 2.1947674418604654e-05, "loss": 0.0679, "step": 710 }, { "epoch": 16.53488372093023, "grad_norm": 4.328090190887451, "learning_rate": 2.180232558139535e-05, "loss": 0.0336, "step": 711 }, { "epoch": 16.558139534883722, "grad_norm": 3.516861915588379, "learning_rate": 2.1656976744186048e-05, "loss": 0.1004, "step": 712 }, { "epoch": 16.58139534883721, "grad_norm": 0.8900988698005676, "learning_rate": 2.1511627906976744e-05, "loss": 0.0083, "step": 713 }, { "epoch": 16.6046511627907, "grad_norm": 4.924454689025879, "learning_rate": 2.1366279069767442e-05, "loss": 0.1108, "step": 714 }, { "epoch": 16.627906976744185, "grad_norm": 4.6815619468688965, "learning_rate": 2.1220930232558138e-05, "loss": 0.0611, "step": 715 }, { "epoch": 16.651162790697676, "grad_norm": 7.625583171844482, "learning_rate": 2.107558139534884e-05, "loss": 0.0698, "step": 716 }, { "epoch": 16.674418604651162, "grad_norm": 2.8963229656219482, "learning_rate": 2.0930232558139536e-05, "loss": 0.0226, "step": 717 }, { "epoch": 16.697674418604652, "grad_norm": 1.349528193473816, "learning_rate": 2.0784883720930235e-05, "loss": 0.0074, "step": 718 }, { "epoch": 16.72093023255814, "grad_norm": 0.9660943746566772, "learning_rate": 2.0639534883720933e-05, "loss": 0.0042, "step": 719 }, { "epoch": 16.74418604651163, "grad_norm": 0.205625519156456, "learning_rate": 2.049418604651163e-05, "loss": 0.0036, "step": 720 }, { "epoch": 16.74418604651163, "eval_accuracy": 0.9081967213114754, "eval_f1": 0.9080296372878436, "eval_loss": 0.3279436528682709, "eval_precision": 0.9080296372878436, "eval_recall": 0.9080296372878436, "eval_runtime": 0.0765, "eval_samples_per_second": 3988.597, "eval_steps_per_second": 65.387, "step": 720 }, { "epoch": 16.767441860465116, "grad_norm": 9.84315299987793, "learning_rate": 2.0348837209302328e-05, "loss": 0.1307, "step": 721 }, { "epoch": 16.790697674418606, "grad_norm": 11.633748054504395, "learning_rate": 2.0203488372093023e-05, "loss": 0.065, "step": 722 }, { "epoch": 16.813953488372093, "grad_norm": 2.6331350803375244, "learning_rate": 2.0058139534883722e-05, "loss": 0.0336, "step": 723 }, { "epoch": 16.837209302325583, "grad_norm": 6.747828006744385, "learning_rate": 1.9912790697674418e-05, "loss": 0.0478, "step": 724 }, { "epoch": 16.86046511627907, "grad_norm": 1.7129838466644287, "learning_rate": 1.9767441860465116e-05, "loss": 0.0403, "step": 725 }, { "epoch": 16.88372093023256, "grad_norm": 1.0863773822784424, "learning_rate": 1.9622093023255815e-05, "loss": 0.0105, "step": 726 }, { "epoch": 16.906976744186046, "grad_norm": 1.2210698127746582, "learning_rate": 1.9476744186046514e-05, "loss": 0.0114, "step": 727 }, { "epoch": 16.930232558139537, "grad_norm": 3.351280689239502, "learning_rate": 1.933139534883721e-05, "loss": 0.0353, "step": 728 }, { "epoch": 16.953488372093023, "grad_norm": 1.9887293577194214, "learning_rate": 1.918604651162791e-05, "loss": 0.0208, "step": 729 }, { "epoch": 16.97674418604651, "grad_norm": 0.7091693878173828, "learning_rate": 1.9040697674418607e-05, "loss": 0.0074, "step": 730 }, { "epoch": 17.0, "grad_norm": 0.6858041882514954, "learning_rate": 1.8895348837209303e-05, "loss": 0.0051, "step": 731 }, { "epoch": 17.023255813953487, "grad_norm": 4.898313045501709, "learning_rate": 1.8750000000000002e-05, "loss": 0.1197, "step": 732 }, { "epoch": 17.046511627906977, "grad_norm": 2.349663496017456, "learning_rate": 1.8604651162790697e-05, "loss": 0.0244, "step": 733 }, { "epoch": 17.069767441860463, "grad_norm": 2.8203916549682617, "learning_rate": 1.8459302325581396e-05, "loss": 0.0801, "step": 734 }, { "epoch": 17.093023255813954, "grad_norm": 4.241973400115967, "learning_rate": 1.831395348837209e-05, "loss": 0.0303, "step": 735 }, { "epoch": 17.11627906976744, "grad_norm": 7.765374660491943, "learning_rate": 1.816860465116279e-05, "loss": 0.0758, "step": 736 }, { "epoch": 17.13953488372093, "grad_norm": 7.214727401733398, "learning_rate": 1.802325581395349e-05, "loss": 0.0461, "step": 737 }, { "epoch": 17.162790697674417, "grad_norm": 7.209427833557129, "learning_rate": 1.7877906976744188e-05, "loss": 0.0354, "step": 738 }, { "epoch": 17.186046511627907, "grad_norm": 4.5369744300842285, "learning_rate": 1.7732558139534887e-05, "loss": 0.0223, "step": 739 }, { "epoch": 17.209302325581394, "grad_norm": 0.8520297408103943, "learning_rate": 1.7587209302325583e-05, "loss": 0.0115, "step": 740 }, { "epoch": 17.209302325581394, "eval_accuracy": 0.8950819672131147, "eval_f1": 0.8945822172297591, "eval_loss": 0.35655906796455383, "eval_precision": 0.8966414996094767, "eval_recall": 0.8937710002584647, "eval_runtime": 0.0741, "eval_samples_per_second": 4118.417, "eval_steps_per_second": 67.515, "step": 740 }, { "epoch": 17.232558139534884, "grad_norm": 2.845405101776123, "learning_rate": 1.744186046511628e-05, "loss": 0.0416, "step": 741 }, { "epoch": 17.25581395348837, "grad_norm": 12.433396339416504, "learning_rate": 1.7296511627906977e-05, "loss": 0.0335, "step": 742 }, { "epoch": 17.27906976744186, "grad_norm": 0.4390711486339569, "learning_rate": 1.7151162790697676e-05, "loss": 0.0062, "step": 743 }, { "epoch": 17.302325581395348, "grad_norm": 2.333735942840576, "learning_rate": 1.700581395348837e-05, "loss": 0.0303, "step": 744 }, { "epoch": 17.325581395348838, "grad_norm": 0.1029660701751709, "learning_rate": 1.686046511627907e-05, "loss": 0.0023, "step": 745 }, { "epoch": 17.348837209302324, "grad_norm": 4.543797016143799, "learning_rate": 1.671511627906977e-05, "loss": 0.0634, "step": 746 }, { "epoch": 17.372093023255815, "grad_norm": 4.657381534576416, "learning_rate": 1.6569767441860464e-05, "loss": 0.0199, "step": 747 }, { "epoch": 17.3953488372093, "grad_norm": 6.3823018074035645, "learning_rate": 1.6424418604651163e-05, "loss": 0.1186, "step": 748 }, { "epoch": 17.41860465116279, "grad_norm": 1.975292682647705, "learning_rate": 1.6279069767441862e-05, "loss": 0.0948, "step": 749 }, { "epoch": 17.441860465116278, "grad_norm": 4.323884010314941, "learning_rate": 1.613372093023256e-05, "loss": 0.1327, "step": 750 }, { "epoch": 17.46511627906977, "grad_norm": 1.2401987314224243, "learning_rate": 1.5988372093023257e-05, "loss": 0.0251, "step": 751 }, { "epoch": 17.488372093023255, "grad_norm": 0.8695264458656311, "learning_rate": 1.5843023255813955e-05, "loss": 0.005, "step": 752 }, { "epoch": 17.511627906976745, "grad_norm": 2.0256264209747314, "learning_rate": 1.569767441860465e-05, "loss": 0.0133, "step": 753 }, { "epoch": 17.53488372093023, "grad_norm": 0.6134868264198303, "learning_rate": 1.555232558139535e-05, "loss": 0.0103, "step": 754 }, { "epoch": 17.558139534883722, "grad_norm": 4.545815467834473, "learning_rate": 1.5406976744186045e-05, "loss": 0.0191, "step": 755 }, { "epoch": 17.58139534883721, "grad_norm": 3.8483781814575195, "learning_rate": 1.5261627906976744e-05, "loss": 0.0311, "step": 756 }, { "epoch": 17.6046511627907, "grad_norm": 2.32814359664917, "learning_rate": 1.5116279069767441e-05, "loss": 0.023, "step": 757 }, { "epoch": 17.627906976744185, "grad_norm": 0.3794184625148773, "learning_rate": 1.4970930232558138e-05, "loss": 0.0054, "step": 758 }, { "epoch": 17.651162790697676, "grad_norm": 1.3837960958480835, "learning_rate": 1.4825581395348839e-05, "loss": 0.0342, "step": 759 }, { "epoch": 17.674418604651162, "grad_norm": 1.8951282501220703, "learning_rate": 1.4680232558139536e-05, "loss": 0.0618, "step": 760 }, { "epoch": 17.674418604651162, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8980054152598138, "eval_loss": 0.3718366026878357, "eval_precision": 0.8990226604393703, "eval_recall": 0.8974756612389075, "eval_runtime": 0.0766, "eval_samples_per_second": 3981.633, "eval_steps_per_second": 65.273, "step": 760 }, { "epoch": 17.697674418604652, "grad_norm": 1.703765630722046, "learning_rate": 1.4534883720930233e-05, "loss": 0.0089, "step": 761 }, { "epoch": 17.72093023255814, "grad_norm": 2.194350004196167, "learning_rate": 1.4389534883720932e-05, "loss": 0.0156, "step": 762 }, { "epoch": 17.74418604651163, "grad_norm": 1.5929670333862305, "learning_rate": 1.424418604651163e-05, "loss": 0.0093, "step": 763 }, { "epoch": 17.767441860465116, "grad_norm": 1.1527395248413086, "learning_rate": 1.4098837209302327e-05, "loss": 0.0087, "step": 764 }, { "epoch": 17.790697674418606, "grad_norm": 0.9619439244270325, "learning_rate": 1.3953488372093024e-05, "loss": 0.0101, "step": 765 }, { "epoch": 17.813953488372093, "grad_norm": 2.041372537612915, "learning_rate": 1.3808139534883721e-05, "loss": 0.0259, "step": 766 }, { "epoch": 17.837209302325583, "grad_norm": 4.107051849365234, "learning_rate": 1.3662790697674418e-05, "loss": 0.0195, "step": 767 }, { "epoch": 17.86046511627907, "grad_norm": 0.8670945763587952, "learning_rate": 1.3517441860465115e-05, "loss": 0.0076, "step": 768 }, { "epoch": 17.88372093023256, "grad_norm": 2.5873591899871826, "learning_rate": 1.3372093023255814e-05, "loss": 0.0325, "step": 769 }, { "epoch": 17.906976744186046, "grad_norm": 3.630141496658325, "learning_rate": 1.3226744186046513e-05, "loss": 0.0261, "step": 770 }, { "epoch": 17.930232558139537, "grad_norm": 0.1693820059299469, "learning_rate": 1.308139534883721e-05, "loss": 0.0036, "step": 771 }, { "epoch": 17.953488372093023, "grad_norm": 0.5527888536453247, "learning_rate": 1.2936046511627909e-05, "loss": 0.0051, "step": 772 }, { "epoch": 17.97674418604651, "grad_norm": 0.8153883814811707, "learning_rate": 1.2790697674418606e-05, "loss": 0.0095, "step": 773 }, { "epoch": 18.0, "grad_norm": 0.2584695518016815, "learning_rate": 1.2645348837209303e-05, "loss": 0.0036, "step": 774 }, { "epoch": 18.023255813953487, "grad_norm": 0.5313040018081665, "learning_rate": 1.25e-05, "loss": 0.0071, "step": 775 }, { "epoch": 18.046511627906977, "grad_norm": 1.0279793739318848, "learning_rate": 1.2354651162790698e-05, "loss": 0.01, "step": 776 }, { "epoch": 18.069767441860463, "grad_norm": 1.0528727769851685, "learning_rate": 1.2209302325581395e-05, "loss": 0.0168, "step": 777 }, { "epoch": 18.093023255813954, "grad_norm": 1.2885291576385498, "learning_rate": 1.2063953488372094e-05, "loss": 0.0077, "step": 778 }, { "epoch": 18.11627906976744, "grad_norm": 0.3943243622779846, "learning_rate": 1.1918604651162791e-05, "loss": 0.0026, "step": 779 }, { "epoch": 18.13953488372093, "grad_norm": 1.1616085767745972, "learning_rate": 1.177325581395349e-05, "loss": 0.0061, "step": 780 }, { "epoch": 18.13953488372093, "eval_accuracy": 0.898360655737705, "eval_f1": 0.8981460534962133, "eval_loss": 0.37966448068618774, "eval_precision": 0.8982758620689655, "eval_recall": 0.8980356681313001, "eval_runtime": 0.0743, "eval_samples_per_second": 4106.506, "eval_steps_per_second": 67.32, "step": 780 }, { "epoch": 18.162790697674417, "grad_norm": 0.11722904443740845, "learning_rate": 1.1627906976744187e-05, "loss": 0.0023, "step": 781 }, { "epoch": 18.186046511627907, "grad_norm": 1.2253535985946655, "learning_rate": 1.1482558139534884e-05, "loss": 0.0182, "step": 782 }, { "epoch": 18.209302325581394, "grad_norm": 0.4131571054458618, "learning_rate": 1.1337209302325581e-05, "loss": 0.0038, "step": 783 }, { "epoch": 18.232558139534884, "grad_norm": 0.5466241240501404, "learning_rate": 1.119186046511628e-05, "loss": 0.0033, "step": 784 }, { "epoch": 18.25581395348837, "grad_norm": 0.8692965507507324, "learning_rate": 1.1046511627906977e-05, "loss": 0.0079, "step": 785 }, { "epoch": 18.27906976744186, "grad_norm": 3.9168105125427246, "learning_rate": 1.0901162790697675e-05, "loss": 0.0861, "step": 786 }, { "epoch": 18.302325581395348, "grad_norm": 1.2862012386322021, "learning_rate": 1.0755813953488372e-05, "loss": 0.0081, "step": 787 }, { "epoch": 18.325581395348838, "grad_norm": 6.316368103027344, "learning_rate": 1.0610465116279069e-05, "loss": 0.044, "step": 788 }, { "epoch": 18.348837209302324, "grad_norm": 7.892980098724365, "learning_rate": 1.0465116279069768e-05, "loss": 0.1037, "step": 789 }, { "epoch": 18.372093023255815, "grad_norm": 7.4783172607421875, "learning_rate": 1.0319767441860467e-05, "loss": 0.0339, "step": 790 }, { "epoch": 18.3953488372093, "grad_norm": 0.7634483575820923, "learning_rate": 1.0174418604651164e-05, "loss": 0.0039, "step": 791 }, { "epoch": 18.41860465116279, "grad_norm": 3.145096778869629, "learning_rate": 1.0029069767441861e-05, "loss": 0.0739, "step": 792 }, { "epoch": 18.441860465116278, "grad_norm": 6.179558277130127, "learning_rate": 9.883720930232558e-06, "loss": 0.0333, "step": 793 }, { "epoch": 18.46511627906977, "grad_norm": 0.10819542407989502, "learning_rate": 9.738372093023257e-06, "loss": 0.0023, "step": 794 }, { "epoch": 18.488372093023255, "grad_norm": 11.038155555725098, "learning_rate": 9.593023255813954e-06, "loss": 0.0816, "step": 795 }, { "epoch": 18.511627906976745, "grad_norm": 4.89321756362915, "learning_rate": 9.447674418604651e-06, "loss": 0.0669, "step": 796 }, { "epoch": 18.53488372093023, "grad_norm": 0.6758410930633545, "learning_rate": 9.302325581395349e-06, "loss": 0.0124, "step": 797 }, { "epoch": 18.558139534883722, "grad_norm": 2.737079381942749, "learning_rate": 9.156976744186046e-06, "loss": 0.0134, "step": 798 }, { "epoch": 18.58139534883721, "grad_norm": 5.82487154006958, "learning_rate": 9.011627906976745e-06, "loss": 0.0826, "step": 799 }, { "epoch": 18.6046511627907, "grad_norm": 0.12569262087345123, "learning_rate": 8.866279069767444e-06, "loss": 0.0021, "step": 800 }, { "epoch": 18.6046511627907, "eval_accuracy": 0.8950819672131147, "eval_f1": 0.8945822172297591, "eval_loss": 0.41386428475379944, "eval_precision": 0.8966414996094767, "eval_recall": 0.8937710002584647, "eval_runtime": 0.074, "eval_samples_per_second": 4121.655, "eval_steps_per_second": 67.568, "step": 800 }, { "epoch": 18.627906976744185, "grad_norm": 7.690613269805908, "learning_rate": 8.72093023255814e-06, "loss": 0.1334, "step": 801 }, { "epoch": 18.651162790697676, "grad_norm": 7.942806720733643, "learning_rate": 8.575581395348838e-06, "loss": 0.0349, "step": 802 }, { "epoch": 18.674418604651162, "grad_norm": 1.4096626043319702, "learning_rate": 8.430232558139535e-06, "loss": 0.0086, "step": 803 }, { "epoch": 18.697674418604652, "grad_norm": 2.184845209121704, "learning_rate": 8.284883720930232e-06, "loss": 0.021, "step": 804 }, { "epoch": 18.72093023255814, "grad_norm": 5.123547554016113, "learning_rate": 8.139534883720931e-06, "loss": 0.0714, "step": 805 }, { "epoch": 18.74418604651163, "grad_norm": 2.141044855117798, "learning_rate": 7.994186046511628e-06, "loss": 0.0227, "step": 806 }, { "epoch": 18.767441860465116, "grad_norm": 0.7044129967689514, "learning_rate": 7.848837209302325e-06, "loss": 0.0064, "step": 807 }, { "epoch": 18.790697674418606, "grad_norm": 0.9102310538291931, "learning_rate": 7.703488372093023e-06, "loss": 0.0056, "step": 808 }, { "epoch": 18.813953488372093, "grad_norm": 6.990550518035889, "learning_rate": 7.558139534883721e-06, "loss": 0.0359, "step": 809 }, { "epoch": 18.837209302325583, "grad_norm": 0.36822423338890076, "learning_rate": 7.4127906976744195e-06, "loss": 0.0036, "step": 810 }, { "epoch": 18.86046511627907, "grad_norm": 1.0232497453689575, "learning_rate": 7.267441860465117e-06, "loss": 0.0077, "step": 811 }, { "epoch": 18.88372093023256, "grad_norm": 2.3726754188537598, "learning_rate": 7.122093023255815e-06, "loss": 0.0625, "step": 812 }, { "epoch": 18.906976744186046, "grad_norm": 1.6161715984344482, "learning_rate": 6.976744186046512e-06, "loss": 0.015, "step": 813 }, { "epoch": 18.930232558139537, "grad_norm": 1.230698823928833, "learning_rate": 6.831395348837209e-06, "loss": 0.0595, "step": 814 }, { "epoch": 18.953488372093023, "grad_norm": 4.402613162994385, "learning_rate": 6.686046511627907e-06, "loss": 0.0847, "step": 815 }, { "epoch": 18.97674418604651, "grad_norm": 6.6045966148376465, "learning_rate": 6.540697674418605e-06, "loss": 0.036, "step": 816 }, { "epoch": 19.0, "grad_norm": 0.6431266069412231, "learning_rate": 6.395348837209303e-06, "loss": 0.0043, "step": 817 }, { "epoch": 19.023255813953487, "grad_norm": 0.4433208703994751, "learning_rate": 6.25e-06, "loss": 0.0032, "step": 818 }, { "epoch": 19.046511627906977, "grad_norm": 3.6785125732421875, "learning_rate": 6.1046511627906975e-06, "loss": 0.1244, "step": 819 }, { "epoch": 19.069767441860463, "grad_norm": 2.5291504859924316, "learning_rate": 5.9593023255813955e-06, "loss": 0.0071, "step": 820 }, { "epoch": 19.069767441860463, "eval_accuracy": 0.9049180327868852, "eval_f1": 0.9046556500555143, "eval_loss": 0.3841802775859833, "eval_precision": 0.9051627384960719, "eval_recall": 0.9043249763074007, "eval_runtime": 0.0744, "eval_samples_per_second": 4097.049, "eval_steps_per_second": 67.165, "step": 820 }, { "epoch": 19.093023255813954, "grad_norm": 8.134071350097656, "learning_rate": 5.8139534883720935e-06, "loss": 0.0308, "step": 821 }, { "epoch": 19.11627906976744, "grad_norm": 0.23603279888629913, "learning_rate": 5.668604651162791e-06, "loss": 0.0028, "step": 822 }, { "epoch": 19.13953488372093, "grad_norm": 5.337473392486572, "learning_rate": 5.523255813953489e-06, "loss": 0.034, "step": 823 }, { "epoch": 19.162790697674417, "grad_norm": 4.512674808502197, "learning_rate": 5.377906976744186e-06, "loss": 0.0775, "step": 824 }, { "epoch": 19.186046511627907, "grad_norm": 1.7216310501098633, "learning_rate": 5.232558139534884e-06, "loss": 0.0056, "step": 825 }, { "epoch": 19.209302325581394, "grad_norm": 1.9352085590362549, "learning_rate": 5.087209302325582e-06, "loss": 0.0423, "step": 826 }, { "epoch": 19.232558139534884, "grad_norm": 10.692374229431152, "learning_rate": 4.941860465116279e-06, "loss": 0.0782, "step": 827 }, { "epoch": 19.25581395348837, "grad_norm": 0.10650653392076492, "learning_rate": 4.796511627906977e-06, "loss": 0.0017, "step": 828 }, { "epoch": 19.27906976744186, "grad_norm": 0.14889536798000336, "learning_rate": 4.651162790697674e-06, "loss": 0.0024, "step": 829 }, { "epoch": 19.302325581395348, "grad_norm": 5.617801666259766, "learning_rate": 4.505813953488372e-06, "loss": 0.0121, "step": 830 }, { "epoch": 19.325581395348838, "grad_norm": 2.859057903289795, "learning_rate": 4.36046511627907e-06, "loss": 0.034, "step": 831 }, { "epoch": 19.348837209302324, "grad_norm": 0.27750277519226074, "learning_rate": 4.2151162790697675e-06, "loss": 0.0044, "step": 832 }, { "epoch": 19.372093023255815, "grad_norm": 0.04883375018835068, "learning_rate": 4.0697674418604655e-06, "loss": 0.0018, "step": 833 }, { "epoch": 19.3953488372093, "grad_norm": 1.9595685005187988, "learning_rate": 3.924418604651163e-06, "loss": 0.011, "step": 834 }, { "epoch": 19.41860465116279, "grad_norm": 1.273353099822998, "learning_rate": 3.7790697674418603e-06, "loss": 0.0214, "step": 835 }, { "epoch": 19.441860465116278, "grad_norm": 6.320394039154053, "learning_rate": 3.6337209302325583e-06, "loss": 0.081, "step": 836 }, { "epoch": 19.46511627906977, "grad_norm": 2.93029522895813, "learning_rate": 3.488372093023256e-06, "loss": 0.0256, "step": 837 }, { "epoch": 19.488372093023255, "grad_norm": 0.4745703339576721, "learning_rate": 3.3430232558139535e-06, "loss": 0.0033, "step": 838 }, { "epoch": 19.511627906976745, "grad_norm": 4.16102933883667, "learning_rate": 3.1976744186046516e-06, "loss": 0.0482, "step": 839 }, { "epoch": 19.53488372093023, "grad_norm": 0.2739373743534088, "learning_rate": 3.0523255813953487e-06, "loss": 0.0045, "step": 840 }, { "epoch": 19.53488372093023, "eval_accuracy": 0.9049180327868852, "eval_f1": 0.9046556500555143, "eval_loss": 0.38903769850730896, "eval_precision": 0.9051627384960719, "eval_recall": 0.9043249763074007, "eval_runtime": 0.0754, "eval_samples_per_second": 4042.913, "eval_steps_per_second": 66.277, "step": 840 }, { "epoch": 19.558139534883722, "grad_norm": 0.87278813123703, "learning_rate": 2.9069767441860468e-06, "loss": 0.0116, "step": 841 }, { "epoch": 19.58139534883721, "grad_norm": 0.7536394596099854, "learning_rate": 2.7616279069767444e-06, "loss": 0.0036, "step": 842 }, { "epoch": 19.6046511627907, "grad_norm": 4.64411735534668, "learning_rate": 2.616279069767442e-06, "loss": 0.0712, "step": 843 }, { "epoch": 19.627906976744185, "grad_norm": 0.5402886271476746, "learning_rate": 2.4709302325581396e-06, "loss": 0.0079, "step": 844 }, { "epoch": 19.651162790697676, "grad_norm": 0.591332733631134, "learning_rate": 2.325581395348837e-06, "loss": 0.0028, "step": 845 }, { "epoch": 19.674418604651162, "grad_norm": 1.511197805404663, "learning_rate": 2.180232558139535e-06, "loss": 0.0216, "step": 846 }, { "epoch": 19.697674418604652, "grad_norm": 2.847646951675415, "learning_rate": 2.0348837209302328e-06, "loss": 0.0246, "step": 847 }, { "epoch": 19.72093023255814, "grad_norm": 0.33959656953811646, "learning_rate": 1.8895348837209302e-06, "loss": 0.0034, "step": 848 }, { "epoch": 19.74418604651163, "grad_norm": 1.8157447576522827, "learning_rate": 1.744186046511628e-06, "loss": 0.0436, "step": 849 }, { "epoch": 19.767441860465116, "grad_norm": 0.3644404113292694, "learning_rate": 1.5988372093023258e-06, "loss": 0.0038, "step": 850 }, { "epoch": 19.790697674418606, "grad_norm": 2.5243899822235107, "learning_rate": 1.4534883720930234e-06, "loss": 0.0115, "step": 851 }, { "epoch": 19.813953488372093, "grad_norm": 5.748671054840088, "learning_rate": 1.308139534883721e-06, "loss": 0.04, "step": 852 }, { "epoch": 19.837209302325583, "grad_norm": 0.5189786553382874, "learning_rate": 1.1627906976744186e-06, "loss": 0.0049, "step": 853 }, { "epoch": 19.86046511627907, "grad_norm": 2.05354380607605, "learning_rate": 1.0174418604651164e-06, "loss": 0.0591, "step": 854 }, { "epoch": 19.88372093023256, "grad_norm": 2.982142925262451, "learning_rate": 8.72093023255814e-07, "loss": 0.0224, "step": 855 }, { "epoch": 19.906976744186046, "grad_norm": 2.0666937828063965, "learning_rate": 7.267441860465117e-07, "loss": 0.0217, "step": 856 }, { "epoch": 19.930232558139537, "grad_norm": 4.0683722496032715, "learning_rate": 5.813953488372093e-07, "loss": 0.108, "step": 857 }, { "epoch": 19.953488372093023, "grad_norm": 10.890898704528809, "learning_rate": 4.36046511627907e-07, "loss": 0.069, "step": 858 }, { "epoch": 19.97674418604651, "grad_norm": 0.7343159914016724, "learning_rate": 2.9069767441860464e-07, "loss": 0.0071, "step": 859 }, { "epoch": 20.0, "grad_norm": 1.6608028411865234, "learning_rate": 1.4534883720930232e-07, "loss": 0.0411, "step": 860 }, { "epoch": 20.0, "eval_accuracy": 0.8950819672131147, "eval_f1": 0.8946732280065612, "eval_loss": 0.3955073356628418, "eval_precision": 0.895995670995671, "eval_recall": 0.894051003704661, "eval_runtime": 0.076, "eval_samples_per_second": 4014.381, "eval_steps_per_second": 65.81, "step": 860 } ], "logging_steps": 1, "max_steps": 860, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 68001814271952.0, "train_batch_size": 64, "trial_name": null, "trial_params": null }