| { | |
| "best_global_step": 400, | |
| "best_metric": 0.9147540983606557, | |
| "best_model_checkpoint": "./modelka_bebro/checkpoint-400", | |
| "epoch": 20.0, | |
| "eval_steps": 20, | |
| "global_step": 860, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.023255813953488372, | |
| "grad_norm": 0.4809555113315582, | |
| "learning_rate": 0.0, | |
| "loss": 0.6913, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.046511627906976744, | |
| "grad_norm": 0.4111059904098511, | |
| "learning_rate": 5.813953488372093e-07, | |
| "loss": 0.6908, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.06976744186046512, | |
| "grad_norm": 0.4677865207195282, | |
| "learning_rate": 1.1627906976744186e-06, | |
| "loss": 0.6945, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.09302325581395349, | |
| "grad_norm": 0.42459365725517273, | |
| "learning_rate": 1.744186046511628e-06, | |
| "loss": 0.6974, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.11627906976744186, | |
| "grad_norm": 0.5455716252326965, | |
| "learning_rate": 2.325581395348837e-06, | |
| "loss": 0.6876, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.13953488372093023, | |
| "grad_norm": 0.3835294842720032, | |
| "learning_rate": 2.9069767441860468e-06, | |
| "loss": 0.6882, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.16279069767441862, | |
| "grad_norm": 0.6538524627685547, | |
| "learning_rate": 3.488372093023256e-06, | |
| "loss": 0.6888, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.18604651162790697, | |
| "grad_norm": 0.40157362818717957, | |
| "learning_rate": 4.0697674418604655e-06, | |
| "loss": 0.6855, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.20930232558139536, | |
| "grad_norm": 0.43001478910446167, | |
| "learning_rate": 4.651162790697674e-06, | |
| "loss": 0.6918, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.23255813953488372, | |
| "grad_norm": 0.5613086819648743, | |
| "learning_rate": 5.232558139534884e-06, | |
| "loss": 0.6977, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.2558139534883721, | |
| "grad_norm": 0.3894985020160675, | |
| "learning_rate": 5.8139534883720935e-06, | |
| "loss": 0.6847, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.27906976744186046, | |
| "grad_norm": 0.4243035912513733, | |
| "learning_rate": 6.395348837209303e-06, | |
| "loss": 0.6934, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.3023255813953488, | |
| "grad_norm": 0.5996053218841553, | |
| "learning_rate": 6.976744186046512e-06, | |
| "loss": 0.6966, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.32558139534883723, | |
| "grad_norm": 0.5445254445075989, | |
| "learning_rate": 7.558139534883721e-06, | |
| "loss": 0.695, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.3488372093023256, | |
| "grad_norm": 0.433915913105011, | |
| "learning_rate": 8.139534883720931e-06, | |
| "loss": 0.6855, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.37209302325581395, | |
| "grad_norm": 0.43263381719589233, | |
| "learning_rate": 8.72093023255814e-06, | |
| "loss": 0.6901, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.3953488372093023, | |
| "grad_norm": 0.5314894318580627, | |
| "learning_rate": 9.302325581395349e-06, | |
| "loss": 0.6966, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.4186046511627907, | |
| "grad_norm": 0.3562123477458954, | |
| "learning_rate": 9.883720930232558e-06, | |
| "loss": 0.6877, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.4418604651162791, | |
| "grad_norm": 0.4751898944377899, | |
| "learning_rate": 1.0465116279069768e-05, | |
| "loss": 0.6834, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.46511627906976744, | |
| "grad_norm": 0.4273492395877838, | |
| "learning_rate": 1.1046511627906977e-05, | |
| "loss": 0.6885, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.46511627906976744, | |
| "eval_accuracy": 0.5901639344262295, | |
| "eval_f1": 0.5314093976229397, | |
| "eval_loss": 0.68656325340271, | |
| "eval_precision": 0.6364057433997221, | |
| "eval_recall": 0.5761178599121219, | |
| "eval_runtime": 0.0751, | |
| "eval_samples_per_second": 4062.686, | |
| "eval_steps_per_second": 66.601, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.4883720930232558, | |
| "grad_norm": 0.5205795764923096, | |
| "learning_rate": 1.1627906976744187e-05, | |
| "loss": 0.6908, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.5116279069767442, | |
| "grad_norm": 0.4466778635978699, | |
| "learning_rate": 1.2209302325581395e-05, | |
| "loss": 0.6861, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.5348837209302325, | |
| "grad_norm": 0.44298067688941956, | |
| "learning_rate": 1.2790697674418606e-05, | |
| "loss": 0.6832, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.5581395348837209, | |
| "grad_norm": 0.39592233300209045, | |
| "learning_rate": 1.3372093023255814e-05, | |
| "loss": 0.6877, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.5813953488372093, | |
| "grad_norm": 0.4810555875301361, | |
| "learning_rate": 1.3953488372093024e-05, | |
| "loss": 0.68, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.6046511627906976, | |
| "grad_norm": 0.47530311346054077, | |
| "learning_rate": 1.4534883720930233e-05, | |
| "loss": 0.6837, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.627906976744186, | |
| "grad_norm": 0.5110817551612854, | |
| "learning_rate": 1.5116279069767441e-05, | |
| "loss": 0.6908, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.6511627906976745, | |
| "grad_norm": 0.44798141717910767, | |
| "learning_rate": 1.569767441860465e-05, | |
| "loss": 0.6756, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.6744186046511628, | |
| "grad_norm": 0.4494743347167969, | |
| "learning_rate": 1.6279069767441862e-05, | |
| "loss": 0.6813, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.6976744186046512, | |
| "grad_norm": 0.4449148178100586, | |
| "learning_rate": 1.686046511627907e-05, | |
| "loss": 0.6786, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.7209302325581395, | |
| "grad_norm": 0.49257639050483704, | |
| "learning_rate": 1.744186046511628e-05, | |
| "loss": 0.6821, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.7441860465116279, | |
| "grad_norm": 0.49307772517204285, | |
| "learning_rate": 1.802325581395349e-05, | |
| "loss": 0.6856, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.7674418604651163, | |
| "grad_norm": 0.4002581536769867, | |
| "learning_rate": 1.8604651162790697e-05, | |
| "loss": 0.6804, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.7906976744186046, | |
| "grad_norm": 0.44210293889045715, | |
| "learning_rate": 1.918604651162791e-05, | |
| "loss": 0.6739, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.813953488372093, | |
| "grad_norm": 0.3651827871799469, | |
| "learning_rate": 1.9767441860465116e-05, | |
| "loss": 0.6839, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.8372093023255814, | |
| "grad_norm": 0.5120950937271118, | |
| "learning_rate": 2.0348837209302328e-05, | |
| "loss": 0.6746, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.8604651162790697, | |
| "grad_norm": 0.4111502766609192, | |
| "learning_rate": 2.0930232558139536e-05, | |
| "loss": 0.6805, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.8837209302325582, | |
| "grad_norm": 0.44530022144317627, | |
| "learning_rate": 2.1511627906976744e-05, | |
| "loss": 0.6833, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.9069767441860465, | |
| "grad_norm": 0.4006333649158478, | |
| "learning_rate": 2.2093023255813955e-05, | |
| "loss": 0.6769, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.9302325581395349, | |
| "grad_norm": 0.4231868386268616, | |
| "learning_rate": 2.2674418604651163e-05, | |
| "loss": 0.6721, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.9302325581395349, | |
| "eval_accuracy": 0.659016393442623, | |
| "eval_f1": 0.649208174820844, | |
| "eval_loss": 0.6683934926986694, | |
| "eval_precision": 0.6947368421052631, | |
| "eval_recall": 0.6673559059188421, | |
| "eval_runtime": 0.0782, | |
| "eval_samples_per_second": 3899.573, | |
| "eval_steps_per_second": 63.927, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.9534883720930233, | |
| "grad_norm": 0.4570966064929962, | |
| "learning_rate": 2.3255813953488374e-05, | |
| "loss": 0.679, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.9767441860465116, | |
| "grad_norm": 0.6141464710235596, | |
| "learning_rate": 2.3837209302325582e-05, | |
| "loss": 0.6768, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.6506465673446655, | |
| "learning_rate": 2.441860465116279e-05, | |
| "loss": 0.6573, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 1.0232558139534884, | |
| "grad_norm": 1.0706639289855957, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.6865, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.0465116279069768, | |
| "grad_norm": 0.4623822867870331, | |
| "learning_rate": 2.5581395348837212e-05, | |
| "loss": 0.6538, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.069767441860465, | |
| "grad_norm": 0.4986269176006317, | |
| "learning_rate": 2.616279069767442e-05, | |
| "loss": 0.6585, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.0930232558139534, | |
| "grad_norm": 0.48053082823753357, | |
| "learning_rate": 2.674418604651163e-05, | |
| "loss": 0.6506, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.1162790697674418, | |
| "grad_norm": 0.5987160801887512, | |
| "learning_rate": 2.7325581395348836e-05, | |
| "loss": 0.6691, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.1395348837209303, | |
| "grad_norm": 0.685001015663147, | |
| "learning_rate": 2.7906976744186048e-05, | |
| "loss": 0.6512, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 1.1627906976744187, | |
| "grad_norm": 0.5840403437614441, | |
| "learning_rate": 2.848837209302326e-05, | |
| "loss": 0.6759, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.1860465116279069, | |
| "grad_norm": 0.6035524606704712, | |
| "learning_rate": 2.9069767441860467e-05, | |
| "loss": 0.6502, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.2093023255813953, | |
| "grad_norm": 0.6236196160316467, | |
| "learning_rate": 2.9651162790697678e-05, | |
| "loss": 0.6575, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.2325581395348837, | |
| "grad_norm": 0.5840880274772644, | |
| "learning_rate": 3.0232558139534883e-05, | |
| "loss": 0.6438, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.255813953488372, | |
| "grad_norm": 0.7563631534576416, | |
| "learning_rate": 3.081395348837209e-05, | |
| "loss": 0.645, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.2790697674418605, | |
| "grad_norm": 0.8720300793647766, | |
| "learning_rate": 3.13953488372093e-05, | |
| "loss": 0.6551, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.302325581395349, | |
| "grad_norm": 0.5763499140739441, | |
| "learning_rate": 3.197674418604651e-05, | |
| "loss": 0.6304, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.3255813953488373, | |
| "grad_norm": 0.5487861037254333, | |
| "learning_rate": 3.2558139534883724e-05, | |
| "loss": 0.6373, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.3488372093023255, | |
| "grad_norm": 0.5714741945266724, | |
| "learning_rate": 3.313953488372093e-05, | |
| "loss": 0.638, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.372093023255814, | |
| "grad_norm": 0.7604700326919556, | |
| "learning_rate": 3.372093023255814e-05, | |
| "loss": 0.6537, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.3953488372093024, | |
| "grad_norm": 0.8128901720046997, | |
| "learning_rate": 3.430232558139535e-05, | |
| "loss": 0.6573, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.3953488372093024, | |
| "eval_accuracy": 0.760655737704918, | |
| "eval_f1": 0.7589037239168805, | |
| "eval_loss": 0.6135660409927368, | |
| "eval_precision": 0.7778153153153153, | |
| "eval_recall": 0.7656801929869905, | |
| "eval_runtime": 0.0732, | |
| "eval_samples_per_second": 4169.193, | |
| "eval_steps_per_second": 68.347, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.4186046511627908, | |
| "grad_norm": 0.6136202216148376, | |
| "learning_rate": 3.488372093023256e-05, | |
| "loss": 0.6257, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.441860465116279, | |
| "grad_norm": 0.6842197179794312, | |
| "learning_rate": 3.5465116279069774e-05, | |
| "loss": 0.6065, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.4651162790697674, | |
| "grad_norm": 0.6682329177856445, | |
| "learning_rate": 3.604651162790698e-05, | |
| "loss": 0.6114, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.4883720930232558, | |
| "grad_norm": 0.7913668155670166, | |
| "learning_rate": 3.662790697674418e-05, | |
| "loss": 0.636, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.5116279069767442, | |
| "grad_norm": 0.7657461762428284, | |
| "learning_rate": 3.7209302325581394e-05, | |
| "loss": 0.6067, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.5348837209302326, | |
| "grad_norm": 0.9346985816955566, | |
| "learning_rate": 3.7790697674418606e-05, | |
| "loss": 0.5902, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.558139534883721, | |
| "grad_norm": 0.7872809767723083, | |
| "learning_rate": 3.837209302325582e-05, | |
| "loss": 0.5731, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.5813953488372094, | |
| "grad_norm": 0.8864495754241943, | |
| "learning_rate": 3.895348837209303e-05, | |
| "loss": 0.581, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.6046511627906976, | |
| "grad_norm": 0.8121451139450073, | |
| "learning_rate": 3.953488372093023e-05, | |
| "loss": 0.5734, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.627906976744186, | |
| "grad_norm": 1.0192168951034546, | |
| "learning_rate": 4.0116279069767444e-05, | |
| "loss": 0.5645, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.6511627906976745, | |
| "grad_norm": 0.9971368312835693, | |
| "learning_rate": 4.0697674418604655e-05, | |
| "loss": 0.5948, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.6744186046511627, | |
| "grad_norm": 0.9530318975448608, | |
| "learning_rate": 4.127906976744187e-05, | |
| "loss": 0.5129, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.697674418604651, | |
| "grad_norm": 0.980096161365509, | |
| "learning_rate": 4.186046511627907e-05, | |
| "loss": 0.5767, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.7209302325581395, | |
| "grad_norm": 0.991642951965332, | |
| "learning_rate": 4.2441860465116276e-05, | |
| "loss": 0.5403, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.744186046511628, | |
| "grad_norm": 0.8262429237365723, | |
| "learning_rate": 4.302325581395349e-05, | |
| "loss": 0.5467, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.7674418604651163, | |
| "grad_norm": 1.1981412172317505, | |
| "learning_rate": 4.36046511627907e-05, | |
| "loss": 0.5081, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.7906976744186047, | |
| "grad_norm": 0.9658230543136597, | |
| "learning_rate": 4.418604651162791e-05, | |
| "loss": 0.5647, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.8139534883720931, | |
| "grad_norm": 1.1805201768875122, | |
| "learning_rate": 4.476744186046512e-05, | |
| "loss": 0.4536, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.8372093023255816, | |
| "grad_norm": 1.101920485496521, | |
| "learning_rate": 4.5348837209302326e-05, | |
| "loss": 0.5361, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.8604651162790697, | |
| "grad_norm": 1.0350321531295776, | |
| "learning_rate": 4.593023255813954e-05, | |
| "loss": 0.5207, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.8604651162790697, | |
| "eval_accuracy": 0.8032786885245902, | |
| "eval_f1": 0.8030224748127099, | |
| "eval_loss": 0.4702507555484772, | |
| "eval_precision": 0.8029135823721811, | |
| "eval_recall": 0.8032006547772896, | |
| "eval_runtime": 0.0743, | |
| "eval_samples_per_second": 4102.555, | |
| "eval_steps_per_second": 67.255, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.8837209302325582, | |
| "grad_norm": 1.0638865232467651, | |
| "learning_rate": 4.651162790697675e-05, | |
| "loss": 0.5351, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.9069767441860463, | |
| "grad_norm": 1.7043132781982422, | |
| "learning_rate": 4.709302325581396e-05, | |
| "loss": 0.4186, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.9302325581395348, | |
| "grad_norm": 1.0048333406448364, | |
| "learning_rate": 4.7674418604651164e-05, | |
| "loss": 0.5134, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.9534883720930232, | |
| "grad_norm": 1.5068817138671875, | |
| "learning_rate": 4.8255813953488375e-05, | |
| "loss": 0.4723, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.9767441860465116, | |
| "grad_norm": 1.1574243307113647, | |
| "learning_rate": 4.883720930232558e-05, | |
| "loss": 0.4904, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 1.2925702333450317, | |
| "learning_rate": 4.941860465116279e-05, | |
| "loss": 0.5023, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 2.0232558139534884, | |
| "grad_norm": 1.679551124572754, | |
| "learning_rate": 5e-05, | |
| "loss": 0.4305, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 2.046511627906977, | |
| "grad_norm": 1.1237919330596924, | |
| "learning_rate": 5.0581395348837214e-05, | |
| "loss": 0.475, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 2.0697674418604652, | |
| "grad_norm": 1.6137198209762573, | |
| "learning_rate": 5.1162790697674425e-05, | |
| "loss": 0.4115, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 2.0930232558139537, | |
| "grad_norm": 1.814559817314148, | |
| "learning_rate": 5.1744186046511636e-05, | |
| "loss": 0.4894, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.116279069767442, | |
| "grad_norm": 2.1933960914611816, | |
| "learning_rate": 5.232558139534884e-05, | |
| "loss": 0.3742, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 2.13953488372093, | |
| "grad_norm": 1.3443443775177002, | |
| "learning_rate": 5.290697674418605e-05, | |
| "loss": 0.396, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.1627906976744184, | |
| "grad_norm": 1.5833871364593506, | |
| "learning_rate": 5.348837209302326e-05, | |
| "loss": 0.3518, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 2.186046511627907, | |
| "grad_norm": 1.5842630863189697, | |
| "learning_rate": 5.406976744186046e-05, | |
| "loss": 0.4174, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.2093023255813953, | |
| "grad_norm": 1.2995549440383911, | |
| "learning_rate": 5.465116279069767e-05, | |
| "loss": 0.3873, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 2.2325581395348837, | |
| "grad_norm": 2.144974946975708, | |
| "learning_rate": 5.5232558139534884e-05, | |
| "loss": 0.4805, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.255813953488372, | |
| "grad_norm": 1.4312434196472168, | |
| "learning_rate": 5.5813953488372095e-05, | |
| "loss": 0.4575, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 2.2790697674418605, | |
| "grad_norm": 2.25107741355896, | |
| "learning_rate": 5.6395348837209306e-05, | |
| "loss": 0.4139, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.302325581395349, | |
| "grad_norm": 1.6673591136932373, | |
| "learning_rate": 5.697674418604652e-05, | |
| "loss": 0.4151, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 2.3255813953488373, | |
| "grad_norm": 1.3179768323898315, | |
| "learning_rate": 5.755813953488373e-05, | |
| "loss": 0.2997, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.3255813953488373, | |
| "eval_accuracy": 0.839344262295082, | |
| "eval_f1": 0.839344262295082, | |
| "eval_loss": 0.3802424669265747, | |
| "eval_precision": 0.8408718876540019, | |
| "eval_recall": 0.8408718876540019, | |
| "eval_runtime": 0.0744, | |
| "eval_samples_per_second": 4101.555, | |
| "eval_steps_per_second": 67.239, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.3488372093023258, | |
| "grad_norm": 2.245741844177246, | |
| "learning_rate": 5.8139534883720933e-05, | |
| "loss": 0.3442, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 2.3720930232558137, | |
| "grad_norm": 2.9141011238098145, | |
| "learning_rate": 5.8720930232558145e-05, | |
| "loss": 0.3489, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.395348837209302, | |
| "grad_norm": 1.829503059387207, | |
| "learning_rate": 5.9302325581395356e-05, | |
| "loss": 0.3544, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.4186046511627906, | |
| "grad_norm": 2.233344793319702, | |
| "learning_rate": 5.9883720930232554e-05, | |
| "loss": 0.3008, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.441860465116279, | |
| "grad_norm": 2.4595718383789062, | |
| "learning_rate": 6.0465116279069765e-05, | |
| "loss": 0.3853, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.4651162790697674, | |
| "grad_norm": 1.539181113243103, | |
| "learning_rate": 6.104651162790698e-05, | |
| "loss": 0.3165, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.488372093023256, | |
| "grad_norm": 1.7545955181121826, | |
| "learning_rate": 6.162790697674418e-05, | |
| "loss": 0.3907, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 2.511627906976744, | |
| "grad_norm": 3.285968065261841, | |
| "learning_rate": 6.22093023255814e-05, | |
| "loss": 0.4645, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.5348837209302326, | |
| "grad_norm": 4.625302314758301, | |
| "learning_rate": 6.27906976744186e-05, | |
| "loss": 0.2975, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 2.558139534883721, | |
| "grad_norm": 1.9037487506866455, | |
| "learning_rate": 6.337209302325582e-05, | |
| "loss": 0.3629, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.5813953488372094, | |
| "grad_norm": 2.768249988555908, | |
| "learning_rate": 6.395348837209303e-05, | |
| "loss": 0.385, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 2.604651162790698, | |
| "grad_norm": 1.8411540985107422, | |
| "learning_rate": 6.453488372093024e-05, | |
| "loss": 0.2562, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 2.6279069767441863, | |
| "grad_norm": 2.3537325859069824, | |
| "learning_rate": 6.511627906976745e-05, | |
| "loss": 0.2486, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 2.6511627906976747, | |
| "grad_norm": 4.455109596252441, | |
| "learning_rate": 6.569767441860465e-05, | |
| "loss": 0.2424, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 2.6744186046511627, | |
| "grad_norm": 3.032252550125122, | |
| "learning_rate": 6.627906976744186e-05, | |
| "loss": 0.3554, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 2.697674418604651, | |
| "grad_norm": 2.142354965209961, | |
| "learning_rate": 6.686046511627908e-05, | |
| "loss": 0.3701, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.7209302325581395, | |
| "grad_norm": 3.0141141414642334, | |
| "learning_rate": 6.744186046511628e-05, | |
| "loss": 0.4602, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 2.744186046511628, | |
| "grad_norm": 2.129042625427246, | |
| "learning_rate": 6.802325581395348e-05, | |
| "loss": 0.4295, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 2.7674418604651163, | |
| "grad_norm": 4.583343029022217, | |
| "learning_rate": 6.86046511627907e-05, | |
| "loss": 0.3599, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 2.7906976744186047, | |
| "grad_norm": 2.799339532852173, | |
| "learning_rate": 6.918604651162791e-05, | |
| "loss": 0.4383, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.7906976744186047, | |
| "eval_accuracy": 0.8426229508196721, | |
| "eval_f1": 0.8426077234024254, | |
| "eval_loss": 0.30933016538619995, | |
| "eval_precision": 0.8432928927895371, | |
| "eval_recall": 0.8437365382958559, | |
| "eval_runtime": 0.0741, | |
| "eval_samples_per_second": 4116.681, | |
| "eval_steps_per_second": 67.487, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.813953488372093, | |
| "grad_norm": 3.329214096069336, | |
| "learning_rate": 6.976744186046513e-05, | |
| "loss": 0.2955, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 2.8372093023255816, | |
| "grad_norm": 1.7641863822937012, | |
| "learning_rate": 7.034883720930233e-05, | |
| "loss": 0.372, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.8604651162790695, | |
| "grad_norm": 2.8948066234588623, | |
| "learning_rate": 7.093023255813955e-05, | |
| "loss": 0.3738, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 2.883720930232558, | |
| "grad_norm": 3.2698655128479004, | |
| "learning_rate": 7.151162790697675e-05, | |
| "loss": 0.4183, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.9069767441860463, | |
| "grad_norm": 2.3984827995300293, | |
| "learning_rate": 7.209302325581396e-05, | |
| "loss": 0.3729, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 2.9302325581395348, | |
| "grad_norm": 2.4366798400878906, | |
| "learning_rate": 7.267441860465116e-05, | |
| "loss": 0.291, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.953488372093023, | |
| "grad_norm": 1.6613179445266724, | |
| "learning_rate": 7.325581395348837e-05, | |
| "loss": 0.2673, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 2.9767441860465116, | |
| "grad_norm": 3.9009366035461426, | |
| "learning_rate": 7.383720930232558e-05, | |
| "loss": 0.3912, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 6.0683417320251465, | |
| "learning_rate": 7.441860465116279e-05, | |
| "loss": 0.3879, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 3.0232558139534884, | |
| "grad_norm": 2.161290407180786, | |
| "learning_rate": 7.500000000000001e-05, | |
| "loss": 0.2878, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 3.046511627906977, | |
| "grad_norm": 5.449894905090332, | |
| "learning_rate": 7.558139534883721e-05, | |
| "loss": 0.3326, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 3.0697674418604652, | |
| "grad_norm": 2.6764752864837646, | |
| "learning_rate": 7.616279069767443e-05, | |
| "loss": 0.2813, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 3.0930232558139537, | |
| "grad_norm": 1.1198005676269531, | |
| "learning_rate": 7.674418604651163e-05, | |
| "loss": 0.2651, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 3.116279069767442, | |
| "grad_norm": 1.7498178482055664, | |
| "learning_rate": 7.732558139534884e-05, | |
| "loss": 0.2171, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 3.13953488372093, | |
| "grad_norm": 4.505956649780273, | |
| "learning_rate": 7.790697674418606e-05, | |
| "loss": 0.2529, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 3.1627906976744184, | |
| "grad_norm": 2.509193181991577, | |
| "learning_rate": 7.848837209302326e-05, | |
| "loss": 0.2102, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 3.186046511627907, | |
| "grad_norm": 1.7064465284347534, | |
| "learning_rate": 7.906976744186047e-05, | |
| "loss": 0.2097, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 3.2093023255813953, | |
| "grad_norm": 1.7659178972244263, | |
| "learning_rate": 7.965116279069767e-05, | |
| "loss": 0.1806, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 3.2325581395348837, | |
| "grad_norm": 3.508608818054199, | |
| "learning_rate": 8.023255813953489e-05, | |
| "loss": 0.1419, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 3.255813953488372, | |
| "grad_norm": 2.2445175647735596, | |
| "learning_rate": 8.081395348837209e-05, | |
| "loss": 0.4201, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 3.255813953488372, | |
| "eval_accuracy": 0.8622950819672132, | |
| "eval_f1": 0.8622225090345896, | |
| "eval_loss": 0.31382298469543457, | |
| "eval_precision": 0.8622291021671826, | |
| "eval_recall": 0.8628844662703541, | |
| "eval_runtime": 0.0729, | |
| "eval_samples_per_second": 4184.713, | |
| "eval_steps_per_second": 68.602, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 3.2790697674418605, | |
| "grad_norm": 2.019263744354248, | |
| "learning_rate": 8.139534883720931e-05, | |
| "loss": 0.2023, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 3.302325581395349, | |
| "grad_norm": 1.6362420320510864, | |
| "learning_rate": 8.197674418604652e-05, | |
| "loss": 0.1729, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 3.3255813953488373, | |
| "grad_norm": 2.1134684085845947, | |
| "learning_rate": 8.255813953488373e-05, | |
| "loss": 0.2688, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 3.3488372093023258, | |
| "grad_norm": 3.522068738937378, | |
| "learning_rate": 8.313953488372094e-05, | |
| "loss": 0.199, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 3.3720930232558137, | |
| "grad_norm": 4.994085788726807, | |
| "learning_rate": 8.372093023255814e-05, | |
| "loss": 0.2641, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 3.395348837209302, | |
| "grad_norm": 2.1998469829559326, | |
| "learning_rate": 8.430232558139536e-05, | |
| "loss": 0.2183, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 3.4186046511627906, | |
| "grad_norm": 1.9816820621490479, | |
| "learning_rate": 8.488372093023255e-05, | |
| "loss": 0.2952, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 3.441860465116279, | |
| "grad_norm": 2.8008265495300293, | |
| "learning_rate": 8.546511627906977e-05, | |
| "loss": 0.2562, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 3.4651162790697674, | |
| "grad_norm": 2.082418918609619, | |
| "learning_rate": 8.604651162790697e-05, | |
| "loss": 0.2626, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 3.488372093023256, | |
| "grad_norm": 3.3041298389434814, | |
| "learning_rate": 8.662790697674419e-05, | |
| "loss": 0.2373, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 3.511627906976744, | |
| "grad_norm": 2.62347674369812, | |
| "learning_rate": 8.72093023255814e-05, | |
| "loss": 0.2749, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 3.5348837209302326, | |
| "grad_norm": 3.0671257972717285, | |
| "learning_rate": 8.779069767441861e-05, | |
| "loss": 0.3538, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 3.558139534883721, | |
| "grad_norm": 2.737697124481201, | |
| "learning_rate": 8.837209302325582e-05, | |
| "loss": 0.3537, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 3.5813953488372094, | |
| "grad_norm": 6.468879699707031, | |
| "learning_rate": 8.895348837209302e-05, | |
| "loss": 0.5041, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 3.604651162790698, | |
| "grad_norm": 3.8154025077819824, | |
| "learning_rate": 8.953488372093024e-05, | |
| "loss": 0.3206, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 3.6279069767441863, | |
| "grad_norm": 1.7028948068618774, | |
| "learning_rate": 9.011627906976745e-05, | |
| "loss": 0.3093, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 3.6511627906976747, | |
| "grad_norm": 2.867501735687256, | |
| "learning_rate": 9.069767441860465e-05, | |
| "loss": 0.2656, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 3.6744186046511627, | |
| "grad_norm": 4.052389621734619, | |
| "learning_rate": 9.127906976744186e-05, | |
| "loss": 0.4386, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 3.697674418604651, | |
| "grad_norm": 1.909553050994873, | |
| "learning_rate": 9.186046511627907e-05, | |
| "loss": 0.2907, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 3.7209302325581395, | |
| "grad_norm": 1.4470328092575073, | |
| "learning_rate": 9.244186046511628e-05, | |
| "loss": 0.1697, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.7209302325581395, | |
| "eval_accuracy": 0.8721311475409836, | |
| "eval_f1": 0.8721256490469893, | |
| "eval_loss": 0.26616957783699036, | |
| "eval_precision": 0.8730848683077983, | |
| "eval_recall": 0.87343844231929, | |
| "eval_runtime": 0.0715, | |
| "eval_samples_per_second": 4265.076, | |
| "eval_steps_per_second": 69.919, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.744186046511628, | |
| "grad_norm": 1.7241488695144653, | |
| "learning_rate": 9.30232558139535e-05, | |
| "loss": 0.248, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 3.7674418604651163, | |
| "grad_norm": 1.370429277420044, | |
| "learning_rate": 9.36046511627907e-05, | |
| "loss": 0.26, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 3.7906976744186047, | |
| "grad_norm": 2.8360958099365234, | |
| "learning_rate": 9.418604651162792e-05, | |
| "loss": 0.3161, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 3.813953488372093, | |
| "grad_norm": 2.840177297592163, | |
| "learning_rate": 9.476744186046512e-05, | |
| "loss": 0.3643, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 3.8372093023255816, | |
| "grad_norm": 1.9618759155273438, | |
| "learning_rate": 9.534883720930233e-05, | |
| "loss": 0.3486, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 3.8604651162790695, | |
| "grad_norm": 3.201247215270996, | |
| "learning_rate": 9.593023255813955e-05, | |
| "loss": 0.293, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 3.883720930232558, | |
| "grad_norm": 3.3960511684417725, | |
| "learning_rate": 9.651162790697675e-05, | |
| "loss": 0.2797, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 3.9069767441860463, | |
| "grad_norm": 2.5416486263275146, | |
| "learning_rate": 9.709302325581396e-05, | |
| "loss": 0.2781, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 3.9302325581395348, | |
| "grad_norm": 4.113210678100586, | |
| "learning_rate": 9.767441860465116e-05, | |
| "loss": 0.2734, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 3.953488372093023, | |
| "grad_norm": 3.773771286010742, | |
| "learning_rate": 9.825581395348838e-05, | |
| "loss": 0.2775, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.9767441860465116, | |
| "grad_norm": 1.3982651233673096, | |
| "learning_rate": 9.883720930232558e-05, | |
| "loss": 0.1802, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 1.2582266330718994, | |
| "learning_rate": 9.94186046511628e-05, | |
| "loss": 0.1667, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 4.023255813953488, | |
| "grad_norm": 6.577229976654053, | |
| "learning_rate": 0.0001, | |
| "loss": 0.4147, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 4.046511627906977, | |
| "grad_norm": 4.923567295074463, | |
| "learning_rate": 9.98546511627907e-05, | |
| "loss": 0.1983, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 4.069767441860465, | |
| "grad_norm": 1.9299342632293701, | |
| "learning_rate": 9.97093023255814e-05, | |
| "loss": 0.2169, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 4.093023255813954, | |
| "grad_norm": 3.2299985885620117, | |
| "learning_rate": 9.95639534883721e-05, | |
| "loss": 0.1698, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 4.116279069767442, | |
| "grad_norm": 2.034163236618042, | |
| "learning_rate": 9.94186046511628e-05, | |
| "loss": 0.191, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 4.1395348837209305, | |
| "grad_norm": 4.542312145233154, | |
| "learning_rate": 9.927325581395349e-05, | |
| "loss": 0.2465, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 4.162790697674419, | |
| "grad_norm": 1.8868296146392822, | |
| "learning_rate": 9.912790697674418e-05, | |
| "loss": 0.2437, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 4.186046511627907, | |
| "grad_norm": 4.527658939361572, | |
| "learning_rate": 9.898255813953488e-05, | |
| "loss": 0.2774, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 4.186046511627907, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8839686525146466, | |
| "eval_loss": 0.3168491721153259, | |
| "eval_precision": 0.8926612403786605, | |
| "eval_recall": 0.8823770138709399, | |
| "eval_runtime": 0.0728, | |
| "eval_samples_per_second": 4190.498, | |
| "eval_steps_per_second": 68.697, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 4.209302325581396, | |
| "grad_norm": 6.06718635559082, | |
| "learning_rate": 9.883720930232558e-05, | |
| "loss": 0.267, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 4.232558139534884, | |
| "grad_norm": 8.161118507385254, | |
| "learning_rate": 9.869186046511628e-05, | |
| "loss": 0.459, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 4.2558139534883725, | |
| "grad_norm": 3.9242475032806396, | |
| "learning_rate": 9.854651162790698e-05, | |
| "loss": 0.2698, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 4.27906976744186, | |
| "grad_norm": 6.657914161682129, | |
| "learning_rate": 9.840116279069768e-05, | |
| "loss": 0.2114, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 4.3023255813953485, | |
| "grad_norm": 2.23410964012146, | |
| "learning_rate": 9.825581395348838e-05, | |
| "loss": 0.1337, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 4.325581395348837, | |
| "grad_norm": 1.6781232357025146, | |
| "learning_rate": 9.811046511627908e-05, | |
| "loss": 0.1269, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 4.348837209302325, | |
| "grad_norm": 1.3690869808197021, | |
| "learning_rate": 9.796511627906976e-05, | |
| "loss": 0.1783, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 4.372093023255814, | |
| "grad_norm": 2.0360560417175293, | |
| "learning_rate": 9.781976744186046e-05, | |
| "loss": 0.281, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 4.395348837209302, | |
| "grad_norm": 3.783336877822876, | |
| "learning_rate": 9.767441860465116e-05, | |
| "loss": 0.3072, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 4.4186046511627906, | |
| "grad_norm": 3.0387606620788574, | |
| "learning_rate": 9.752906976744186e-05, | |
| "loss": 0.2682, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 4.441860465116279, | |
| "grad_norm": 1.6739147901535034, | |
| "learning_rate": 9.738372093023256e-05, | |
| "loss": 0.1597, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 4.465116279069767, | |
| "grad_norm": 4.560539245605469, | |
| "learning_rate": 9.723837209302326e-05, | |
| "loss": 0.2328, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 4.488372093023256, | |
| "grad_norm": 2.564955711364746, | |
| "learning_rate": 9.709302325581396e-05, | |
| "loss": 0.2731, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 4.511627906976744, | |
| "grad_norm": 6.892247676849365, | |
| "learning_rate": 9.694767441860465e-05, | |
| "loss": 0.2358, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 4.534883720930233, | |
| "grad_norm": 2.4123950004577637, | |
| "learning_rate": 9.680232558139535e-05, | |
| "loss": 0.2267, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 4.558139534883721, | |
| "grad_norm": 2.6707470417022705, | |
| "learning_rate": 9.665697674418605e-05, | |
| "loss": 0.1415, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 4.5813953488372094, | |
| "grad_norm": 1.8982809782028198, | |
| "learning_rate": 9.651162790697675e-05, | |
| "loss": 0.2119, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 4.604651162790698, | |
| "grad_norm": 3.8330225944519043, | |
| "learning_rate": 9.636627906976745e-05, | |
| "loss": 0.152, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 4.627906976744186, | |
| "grad_norm": 2.504080057144165, | |
| "learning_rate": 9.622093023255815e-05, | |
| "loss": 0.1623, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 4.651162790697675, | |
| "grad_norm": 1.6243199110031128, | |
| "learning_rate": 9.607558139534885e-05, | |
| "loss": 0.1537, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 4.651162790697675, | |
| "eval_accuracy": 0.8950819672131147, | |
| "eval_f1": 0.8942532720811303, | |
| "eval_loss": 0.2364664524793625, | |
| "eval_precision": 0.8994351024697845, | |
| "eval_recall": 0.892930989919876, | |
| "eval_runtime": 0.0733, | |
| "eval_samples_per_second": 4158.906, | |
| "eval_steps_per_second": 68.179, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 4.674418604651163, | |
| "grad_norm": 3.362497091293335, | |
| "learning_rate": 9.593023255813955e-05, | |
| "loss": 0.2097, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 4.6976744186046515, | |
| "grad_norm": 1.36155366897583, | |
| "learning_rate": 9.578488372093024e-05, | |
| "loss": 0.134, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 4.720930232558139, | |
| "grad_norm": 2.335468053817749, | |
| "learning_rate": 9.563953488372094e-05, | |
| "loss": 0.2128, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 4.7441860465116275, | |
| "grad_norm": 2.431711196899414, | |
| "learning_rate": 9.549418604651163e-05, | |
| "loss": 0.2257, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 4.767441860465116, | |
| "grad_norm": 2.425615072250366, | |
| "learning_rate": 9.534883720930233e-05, | |
| "loss": 0.1691, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 4.790697674418604, | |
| "grad_norm": 3.3857626914978027, | |
| "learning_rate": 9.520348837209303e-05, | |
| "loss": 0.2282, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 4.813953488372093, | |
| "grad_norm": 2.252249240875244, | |
| "learning_rate": 9.505813953488373e-05, | |
| "loss": 0.2333, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 4.837209302325581, | |
| "grad_norm": 4.851521015167236, | |
| "learning_rate": 9.491279069767442e-05, | |
| "loss": 0.2621, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 4.8604651162790695, | |
| "grad_norm": 1.249811053276062, | |
| "learning_rate": 9.476744186046512e-05, | |
| "loss": 0.075, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 4.883720930232558, | |
| "grad_norm": 1.6059314012527466, | |
| "learning_rate": 9.462209302325582e-05, | |
| "loss": 0.1878, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.906976744186046, | |
| "grad_norm": 7.576606750488281, | |
| "learning_rate": 9.447674418604652e-05, | |
| "loss": 0.2258, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 4.930232558139535, | |
| "grad_norm": 2.4628143310546875, | |
| "learning_rate": 9.433139534883722e-05, | |
| "loss": 0.1463, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 4.953488372093023, | |
| "grad_norm": 4.781926155090332, | |
| "learning_rate": 9.418604651162792e-05, | |
| "loss": 0.2608, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 4.976744186046512, | |
| "grad_norm": 3.2314810752868652, | |
| "learning_rate": 9.40406976744186e-05, | |
| "loss": 0.2344, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 3.6826560497283936, | |
| "learning_rate": 9.38953488372093e-05, | |
| "loss": 0.17, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 5.023255813953488, | |
| "grad_norm": 1.804560661315918, | |
| "learning_rate": 9.375e-05, | |
| "loss": 0.1241, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 5.046511627906977, | |
| "grad_norm": 3.6199300289154053, | |
| "learning_rate": 9.36046511627907e-05, | |
| "loss": 0.1335, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 5.069767441860465, | |
| "grad_norm": 2.37376070022583, | |
| "learning_rate": 9.34593023255814e-05, | |
| "loss": 0.1883, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 5.093023255813954, | |
| "grad_norm": 4.643337726593018, | |
| "learning_rate": 9.33139534883721e-05, | |
| "loss": 0.187, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 5.116279069767442, | |
| "grad_norm": 6.206487655639648, | |
| "learning_rate": 9.31686046511628e-05, | |
| "loss": 0.166, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 5.116279069767442, | |
| "eval_accuracy": 0.8819672131147541, | |
| "eval_f1": 0.8815073815073815, | |
| "eval_loss": 0.27603965997695923, | |
| "eval_precision": 0.8827922077922078, | |
| "eval_recall": 0.8809123804600671, | |
| "eval_runtime": 0.0725, | |
| "eval_samples_per_second": 4205.887, | |
| "eval_steps_per_second": 68.949, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 5.1395348837209305, | |
| "grad_norm": 2.4301323890686035, | |
| "learning_rate": 9.30232558139535e-05, | |
| "loss": 0.0946, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 5.162790697674419, | |
| "grad_norm": 4.870151519775391, | |
| "learning_rate": 9.28779069767442e-05, | |
| "loss": 0.2674, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 5.186046511627907, | |
| "grad_norm": 5.952517032623291, | |
| "learning_rate": 9.273255813953488e-05, | |
| "loss": 0.2605, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 5.209302325581396, | |
| "grad_norm": 1.8462685346603394, | |
| "learning_rate": 9.258720930232558e-05, | |
| "loss": 0.1902, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 5.232558139534884, | |
| "grad_norm": 2.608715295791626, | |
| "learning_rate": 9.244186046511628e-05, | |
| "loss": 0.1392, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 5.2558139534883725, | |
| "grad_norm": 4.296781539916992, | |
| "learning_rate": 9.229651162790698e-05, | |
| "loss": 0.2599, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 5.27906976744186, | |
| "grad_norm": 6.4339165687561035, | |
| "learning_rate": 9.215116279069768e-05, | |
| "loss": 0.162, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 5.3023255813953485, | |
| "grad_norm": 4.432989597320557, | |
| "learning_rate": 9.200581395348837e-05, | |
| "loss": 0.1832, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 5.325581395348837, | |
| "grad_norm": 1.9655786752700806, | |
| "learning_rate": 9.186046511627907e-05, | |
| "loss": 0.1192, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 5.348837209302325, | |
| "grad_norm": 4.3318891525268555, | |
| "learning_rate": 9.171511627906977e-05, | |
| "loss": 0.244, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 5.372093023255814, | |
| "grad_norm": 2.993924856185913, | |
| "learning_rate": 9.156976744186047e-05, | |
| "loss": 0.1828, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 5.395348837209302, | |
| "grad_norm": 2.020063638687134, | |
| "learning_rate": 9.142441860465116e-05, | |
| "loss": 0.1642, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 5.4186046511627906, | |
| "grad_norm": 2.384181499481201, | |
| "learning_rate": 9.127906976744186e-05, | |
| "loss": 0.1752, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 5.441860465116279, | |
| "grad_norm": 1.9194653034210205, | |
| "learning_rate": 9.113372093023255e-05, | |
| "loss": 0.0996, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 5.465116279069767, | |
| "grad_norm": 3.680722713470459, | |
| "learning_rate": 9.098837209302325e-05, | |
| "loss": 0.1925, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 5.488372093023256, | |
| "grad_norm": 3.332195520401001, | |
| "learning_rate": 9.084302325581395e-05, | |
| "loss": 0.13, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 5.511627906976744, | |
| "grad_norm": 2.6394619941711426, | |
| "learning_rate": 9.069767441860465e-05, | |
| "loss": 0.193, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 5.534883720930233, | |
| "grad_norm": 4.5959038734436035, | |
| "learning_rate": 9.055232558139536e-05, | |
| "loss": 0.2303, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 5.558139534883721, | |
| "grad_norm": 6.544390678405762, | |
| "learning_rate": 9.040697674418606e-05, | |
| "loss": 0.2432, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 5.5813953488372094, | |
| "grad_norm": 1.6620129346847534, | |
| "learning_rate": 9.026162790697675e-05, | |
| "loss": 0.1378, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 5.5813953488372094, | |
| "eval_accuracy": 0.8819672131147541, | |
| "eval_f1": 0.8816810344827587, | |
| "eval_loss": 0.21341156959533691, | |
| "eval_precision": 0.8819660110420979, | |
| "eval_recall": 0.8814723873524597, | |
| "eval_runtime": 0.091, | |
| "eval_samples_per_second": 3353.025, | |
| "eval_steps_per_second": 54.968, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 5.604651162790698, | |
| "grad_norm": 1.251617431640625, | |
| "learning_rate": 9.011627906976745e-05, | |
| "loss": 0.1108, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 5.627906976744186, | |
| "grad_norm": 2.331637144088745, | |
| "learning_rate": 8.997093023255815e-05, | |
| "loss": 0.1591, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 5.651162790697675, | |
| "grad_norm": 2.126330852508545, | |
| "learning_rate": 8.982558139534884e-05, | |
| "loss": 0.1115, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 5.674418604651163, | |
| "grad_norm": 2.0312886238098145, | |
| "learning_rate": 8.968023255813954e-05, | |
| "loss": 0.1009, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 5.6976744186046515, | |
| "grad_norm": 3.0067551136016846, | |
| "learning_rate": 8.953488372093024e-05, | |
| "loss": 0.1851, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 5.720930232558139, | |
| "grad_norm": 3.1301369667053223, | |
| "learning_rate": 8.938953488372094e-05, | |
| "loss": 0.107, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 5.7441860465116275, | |
| "grad_norm": 6.043979167938232, | |
| "learning_rate": 8.924418604651164e-05, | |
| "loss": 0.273, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 5.767441860465116, | |
| "grad_norm": 3.7104833126068115, | |
| "learning_rate": 8.909883720930234e-05, | |
| "loss": 0.1299, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 5.790697674418604, | |
| "grad_norm": 2.3856964111328125, | |
| "learning_rate": 8.895348837209302e-05, | |
| "loss": 0.1252, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 5.813953488372093, | |
| "grad_norm": 5.702451705932617, | |
| "learning_rate": 8.880813953488372e-05, | |
| "loss": 0.1939, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 5.837209302325581, | |
| "grad_norm": 3.5763566493988037, | |
| "learning_rate": 8.866279069767442e-05, | |
| "loss": 0.2426, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 5.8604651162790695, | |
| "grad_norm": 4.04267692565918, | |
| "learning_rate": 8.851744186046512e-05, | |
| "loss": 0.1966, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 5.883720930232558, | |
| "grad_norm": 2.391784191131592, | |
| "learning_rate": 8.837209302325582e-05, | |
| "loss": 0.2284, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 5.906976744186046, | |
| "grad_norm": 4.925626754760742, | |
| "learning_rate": 8.822674418604652e-05, | |
| "loss": 0.1559, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 5.930232558139535, | |
| "grad_norm": 5.659231185913086, | |
| "learning_rate": 8.808139534883722e-05, | |
| "loss": 0.1527, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 5.953488372093023, | |
| "grad_norm": 1.3519715070724487, | |
| "learning_rate": 8.793604651162792e-05, | |
| "loss": 0.097, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 5.976744186046512, | |
| "grad_norm": 3.517763137817383, | |
| "learning_rate": 8.779069767441861e-05, | |
| "loss": 0.2737, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 3.0338218212127686, | |
| "learning_rate": 8.76453488372093e-05, | |
| "loss": 0.1154, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 6.023255813953488, | |
| "grad_norm": 2.716304063796997, | |
| "learning_rate": 8.75e-05, | |
| "loss": 0.2143, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 6.046511627906977, | |
| "grad_norm": 1.8077433109283447, | |
| "learning_rate": 8.73546511627907e-05, | |
| "loss": 0.1411, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 6.046511627906977, | |
| "eval_accuracy": 0.8950819672131147, | |
| "eval_f1": 0.8945822172297591, | |
| "eval_loss": 0.23519852757453918, | |
| "eval_precision": 0.8966414996094767, | |
| "eval_recall": 0.8937710002584647, | |
| "eval_runtime": 0.0746, | |
| "eval_samples_per_second": 4089.283, | |
| "eval_steps_per_second": 67.037, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 6.069767441860465, | |
| "grad_norm": 4.399131774902344, | |
| "learning_rate": 8.72093023255814e-05, | |
| "loss": 0.1678, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 6.093023255813954, | |
| "grad_norm": 1.8267767429351807, | |
| "learning_rate": 8.70639534883721e-05, | |
| "loss": 0.0858, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 6.116279069767442, | |
| "grad_norm": 4.386148929595947, | |
| "learning_rate": 8.69186046511628e-05, | |
| "loss": 0.1666, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 6.1395348837209305, | |
| "grad_norm": 4.314615249633789, | |
| "learning_rate": 8.67732558139535e-05, | |
| "loss": 0.1782, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 6.162790697674419, | |
| "grad_norm": 1.9564558267593384, | |
| "learning_rate": 8.662790697674419e-05, | |
| "loss": 0.203, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 6.186046511627907, | |
| "grad_norm": 3.557976484298706, | |
| "learning_rate": 8.648255813953489e-05, | |
| "loss": 0.1347, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 6.209302325581396, | |
| "grad_norm": 2.464057207107544, | |
| "learning_rate": 8.633720930232559e-05, | |
| "loss": 0.0906, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 6.232558139534884, | |
| "grad_norm": 2.893970251083374, | |
| "learning_rate": 8.619186046511628e-05, | |
| "loss": 0.0893, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 6.2558139534883725, | |
| "grad_norm": 4.821974277496338, | |
| "learning_rate": 8.604651162790697e-05, | |
| "loss": 0.1008, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 6.27906976744186, | |
| "grad_norm": 2.767670154571533, | |
| "learning_rate": 8.590116279069767e-05, | |
| "loss": 0.1072, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 6.3023255813953485, | |
| "grad_norm": 2.337888240814209, | |
| "learning_rate": 8.575581395348837e-05, | |
| "loss": 0.1691, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 6.325581395348837, | |
| "grad_norm": 2.4074652194976807, | |
| "learning_rate": 8.561046511627907e-05, | |
| "loss": 0.1671, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 6.348837209302325, | |
| "grad_norm": 6.366968631744385, | |
| "learning_rate": 8.546511627906977e-05, | |
| "loss": 0.1741, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 6.372093023255814, | |
| "grad_norm": 2.4231204986572266, | |
| "learning_rate": 8.531976744186047e-05, | |
| "loss": 0.1563, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 6.395348837209302, | |
| "grad_norm": 3.424386739730835, | |
| "learning_rate": 8.517441860465117e-05, | |
| "loss": 0.2211, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 6.4186046511627906, | |
| "grad_norm": 6.0630717277526855, | |
| "learning_rate": 8.502906976744187e-05, | |
| "loss": 0.3506, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 6.441860465116279, | |
| "grad_norm": 2.5241973400115967, | |
| "learning_rate": 8.488372093023255e-05, | |
| "loss": 0.1239, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 6.465116279069767, | |
| "grad_norm": 1.1857047080993652, | |
| "learning_rate": 8.473837209302325e-05, | |
| "loss": 0.1218, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 6.488372093023256, | |
| "grad_norm": 2.826883316040039, | |
| "learning_rate": 8.459302325581395e-05, | |
| "loss": 0.1296, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 6.511627906976744, | |
| "grad_norm": 3.720126152038574, | |
| "learning_rate": 8.444767441860465e-05, | |
| "loss": 0.1594, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 6.511627906976744, | |
| "eval_accuracy": 0.8819672131147541, | |
| "eval_f1": 0.8815073815073815, | |
| "eval_loss": 0.25715914368629456, | |
| "eval_precision": 0.8827922077922078, | |
| "eval_recall": 0.8809123804600671, | |
| "eval_runtime": 0.0733, | |
| "eval_samples_per_second": 4160.082, | |
| "eval_steps_per_second": 68.198, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 6.534883720930233, | |
| "grad_norm": 2.7126853466033936, | |
| "learning_rate": 8.430232558139536e-05, | |
| "loss": 0.1234, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 6.558139534883721, | |
| "grad_norm": 2.641782522201538, | |
| "learning_rate": 8.415697674418606e-05, | |
| "loss": 0.1515, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 6.5813953488372094, | |
| "grad_norm": 5.715346813201904, | |
| "learning_rate": 8.401162790697676e-05, | |
| "loss": 0.1463, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 6.604651162790698, | |
| "grad_norm": 4.03624153137207, | |
| "learning_rate": 8.386627906976746e-05, | |
| "loss": 0.1425, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 6.627906976744186, | |
| "grad_norm": 0.8614564538002014, | |
| "learning_rate": 8.372093023255814e-05, | |
| "loss": 0.1115, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 6.651162790697675, | |
| "grad_norm": 3.0801596641540527, | |
| "learning_rate": 8.357558139534884e-05, | |
| "loss": 0.1954, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 6.674418604651163, | |
| "grad_norm": 2.3671586513519287, | |
| "learning_rate": 8.343023255813954e-05, | |
| "loss": 0.1223, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 6.6976744186046515, | |
| "grad_norm": 4.468645095825195, | |
| "learning_rate": 8.328488372093024e-05, | |
| "loss": 0.1422, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 6.720930232558139, | |
| "grad_norm": 2.540923833847046, | |
| "learning_rate": 8.313953488372094e-05, | |
| "loss": 0.127, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 6.7441860465116275, | |
| "grad_norm": 1.3170732259750366, | |
| "learning_rate": 8.299418604651164e-05, | |
| "loss": 0.0479, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 6.767441860465116, | |
| "grad_norm": 3.5731353759765625, | |
| "learning_rate": 8.284883720930234e-05, | |
| "loss": 0.1118, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 6.790697674418604, | |
| "grad_norm": 6.928558826446533, | |
| "learning_rate": 8.270348837209303e-05, | |
| "loss": 0.2011, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 6.813953488372093, | |
| "grad_norm": 4.524282932281494, | |
| "learning_rate": 8.255813953488373e-05, | |
| "loss": 0.1252, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 6.837209302325581, | |
| "grad_norm": 7.167698383331299, | |
| "learning_rate": 8.241279069767442e-05, | |
| "loss": 0.153, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 6.8604651162790695, | |
| "grad_norm": 4.981342792510986, | |
| "learning_rate": 8.226744186046512e-05, | |
| "loss": 0.2423, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 6.883720930232558, | |
| "grad_norm": 2.139774799346924, | |
| "learning_rate": 8.212209302325582e-05, | |
| "loss": 0.1548, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 6.906976744186046, | |
| "grad_norm": 2.5026113986968994, | |
| "learning_rate": 8.197674418604652e-05, | |
| "loss": 0.0935, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 6.930232558139535, | |
| "grad_norm": 1.633834719657898, | |
| "learning_rate": 8.183139534883721e-05, | |
| "loss": 0.1224, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 6.953488372093023, | |
| "grad_norm": 4.0517168045043945, | |
| "learning_rate": 8.168604651162791e-05, | |
| "loss": 0.106, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 6.976744186046512, | |
| "grad_norm": 4.622942924499512, | |
| "learning_rate": 8.154069767441861e-05, | |
| "loss": 0.1923, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 6.976744186046512, | |
| "eval_accuracy": 0.8918032786885246, | |
| "eval_f1": 0.8912373974778747, | |
| "eval_loss": 0.276712030172348, | |
| "eval_precision": 0.8937043795620438, | |
| "eval_recall": 0.8903463427242182, | |
| "eval_runtime": 0.0737, | |
| "eval_samples_per_second": 4140.129, | |
| "eval_steps_per_second": 67.871, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "grad_norm": 3.300420045852661, | |
| "learning_rate": 8.139534883720931e-05, | |
| "loss": 0.1218, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 7.023255813953488, | |
| "grad_norm": 2.9694247245788574, | |
| "learning_rate": 8.125000000000001e-05, | |
| "loss": 0.1067, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 7.046511627906977, | |
| "grad_norm": 3.805917978286743, | |
| "learning_rate": 8.11046511627907e-05, | |
| "loss": 0.1165, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 7.069767441860465, | |
| "grad_norm": 2.0280096530914307, | |
| "learning_rate": 8.09593023255814e-05, | |
| "loss": 0.119, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 7.093023255813954, | |
| "grad_norm": 5.049257278442383, | |
| "learning_rate": 8.081395348837209e-05, | |
| "loss": 0.1838, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 7.116279069767442, | |
| "grad_norm": 6.135448455810547, | |
| "learning_rate": 8.066860465116279e-05, | |
| "loss": 0.2193, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 7.1395348837209305, | |
| "grad_norm": 2.0232150554656982, | |
| "learning_rate": 8.052325581395349e-05, | |
| "loss": 0.0829, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 7.162790697674419, | |
| "grad_norm": 3.0096096992492676, | |
| "learning_rate": 8.037790697674419e-05, | |
| "loss": 0.1394, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 7.186046511627907, | |
| "grad_norm": 0.9171739220619202, | |
| "learning_rate": 8.023255813953489e-05, | |
| "loss": 0.1058, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 7.209302325581396, | |
| "grad_norm": 3.929137706756592, | |
| "learning_rate": 8.008720930232559e-05, | |
| "loss": 0.2189, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 7.232558139534884, | |
| "grad_norm": 1.1936514377593994, | |
| "learning_rate": 7.994186046511629e-05, | |
| "loss": 0.0981, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 7.2558139534883725, | |
| "grad_norm": 2.542099714279175, | |
| "learning_rate": 7.979651162790697e-05, | |
| "loss": 0.1052, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 7.27906976744186, | |
| "grad_norm": 2.42411732673645, | |
| "learning_rate": 7.965116279069767e-05, | |
| "loss": 0.0834, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 7.3023255813953485, | |
| "grad_norm": 1.07964289188385, | |
| "learning_rate": 7.950581395348837e-05, | |
| "loss": 0.0684, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 7.325581395348837, | |
| "grad_norm": 0.9596773386001587, | |
| "learning_rate": 7.936046511627907e-05, | |
| "loss": 0.0791, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 7.348837209302325, | |
| "grad_norm": 3.508575201034546, | |
| "learning_rate": 7.921511627906977e-05, | |
| "loss": 0.145, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 7.372093023255814, | |
| "grad_norm": 3.8409645557403564, | |
| "learning_rate": 7.906976744186047e-05, | |
| "loss": 0.1057, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 7.395348837209302, | |
| "grad_norm": 5.247814655303955, | |
| "learning_rate": 7.892441860465116e-05, | |
| "loss": 0.0731, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 7.4186046511627906, | |
| "grad_norm": 2.5410549640655518, | |
| "learning_rate": 7.877906976744186e-05, | |
| "loss": 0.2125, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 7.441860465116279, | |
| "grad_norm": 3.2063095569610596, | |
| "learning_rate": 7.863372093023256e-05, | |
| "loss": 0.0896, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 7.441860465116279, | |
| "eval_accuracy": 0.9114754098360656, | |
| "eval_f1": 0.9114411381991419, | |
| "eval_loss": 0.22706902027130127, | |
| "eval_precision": 0.9115497076023391, | |
| "eval_recall": 0.9122943051606789, | |
| "eval_runtime": 0.0744, | |
| "eval_samples_per_second": 4097.456, | |
| "eval_steps_per_second": 67.171, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 7.465116279069767, | |
| "grad_norm": 1.1547659635543823, | |
| "learning_rate": 7.848837209302326e-05, | |
| "loss": 0.0425, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 7.488372093023256, | |
| "grad_norm": 10.542337417602539, | |
| "learning_rate": 7.834302325581395e-05, | |
| "loss": 0.4056, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 7.511627906976744, | |
| "grad_norm": 6.774454593658447, | |
| "learning_rate": 7.819767441860465e-05, | |
| "loss": 0.106, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 7.534883720930233, | |
| "grad_norm": 3.680190086364746, | |
| "learning_rate": 7.805232558139536e-05, | |
| "loss": 0.1208, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 7.558139534883721, | |
| "grad_norm": 3.1805880069732666, | |
| "learning_rate": 7.790697674418606e-05, | |
| "loss": 0.0968, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 7.5813953488372094, | |
| "grad_norm": 3.887376070022583, | |
| "learning_rate": 7.776162790697676e-05, | |
| "loss": 0.3548, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 7.604651162790698, | |
| "grad_norm": 2.4842185974121094, | |
| "learning_rate": 7.761627906976745e-05, | |
| "loss": 0.0782, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 7.627906976744186, | |
| "grad_norm": 4.861859321594238, | |
| "learning_rate": 7.747093023255815e-05, | |
| "loss": 0.1529, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 7.651162790697675, | |
| "grad_norm": 4.307983875274658, | |
| "learning_rate": 7.732558139534884e-05, | |
| "loss": 0.1938, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 7.674418604651163, | |
| "grad_norm": 6.8119587898254395, | |
| "learning_rate": 7.718023255813954e-05, | |
| "loss": 0.1118, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 7.6976744186046515, | |
| "grad_norm": 1.8230775594711304, | |
| "learning_rate": 7.703488372093024e-05, | |
| "loss": 0.0551, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 7.720930232558139, | |
| "grad_norm": 1.6934735774993896, | |
| "learning_rate": 7.688953488372094e-05, | |
| "loss": 0.0629, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 7.7441860465116275, | |
| "grad_norm": 2.537760019302368, | |
| "learning_rate": 7.674418604651163e-05, | |
| "loss": 0.1084, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 7.767441860465116, | |
| "grad_norm": 1.6460330486297607, | |
| "learning_rate": 7.659883720930233e-05, | |
| "loss": 0.0927, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 7.790697674418604, | |
| "grad_norm": 1.8621909618377686, | |
| "learning_rate": 7.645348837209303e-05, | |
| "loss": 0.107, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 7.813953488372093, | |
| "grad_norm": 7.628468036651611, | |
| "learning_rate": 7.630813953488373e-05, | |
| "loss": 0.1038, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 7.837209302325581, | |
| "grad_norm": 2.3361170291900635, | |
| "learning_rate": 7.616279069767443e-05, | |
| "loss": 0.1017, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 7.8604651162790695, | |
| "grad_norm": 1.8995561599731445, | |
| "learning_rate": 7.601744186046513e-05, | |
| "loss": 0.0414, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 7.883720930232558, | |
| "grad_norm": 5.01551628112793, | |
| "learning_rate": 7.587209302325581e-05, | |
| "loss": 0.1229, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 7.906976744186046, | |
| "grad_norm": 4.123791694641113, | |
| "learning_rate": 7.572674418604651e-05, | |
| "loss": 0.099, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 7.906976744186046, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9016129032258065, | |
| "eval_loss": 0.22697412967681885, | |
| "eval_precision": 0.9018878472520857, | |
| "eval_recall": 0.9025803394503318, | |
| "eval_runtime": 0.0787, | |
| "eval_samples_per_second": 3874.417, | |
| "eval_steps_per_second": 63.515, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 7.930232558139535, | |
| "grad_norm": 1.9877846240997314, | |
| "learning_rate": 7.558139534883721e-05, | |
| "loss": 0.0625, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 7.953488372093023, | |
| "grad_norm": 1.6333461999893188, | |
| "learning_rate": 7.543604651162791e-05, | |
| "loss": 0.051, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 7.976744186046512, | |
| "grad_norm": 8.968517303466797, | |
| "learning_rate": 7.529069767441861e-05, | |
| "loss": 0.0781, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 2.306673765182495, | |
| "learning_rate": 7.514534883720931e-05, | |
| "loss": 0.0684, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 8.023255813953488, | |
| "grad_norm": 5.471800804138184, | |
| "learning_rate": 7.500000000000001e-05, | |
| "loss": 0.1, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 8.046511627906977, | |
| "grad_norm": 6.212913513183594, | |
| "learning_rate": 7.48546511627907e-05, | |
| "loss": 0.0966, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 8.069767441860465, | |
| "grad_norm": 7.509064197540283, | |
| "learning_rate": 7.47093023255814e-05, | |
| "loss": 0.2218, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 8.093023255813954, | |
| "grad_norm": 5.755818843841553, | |
| "learning_rate": 7.456395348837209e-05, | |
| "loss": 0.2037, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 8.116279069767442, | |
| "grad_norm": 2.3190879821777344, | |
| "learning_rate": 7.441860465116279e-05, | |
| "loss": 0.1259, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 8.13953488372093, | |
| "grad_norm": 0.9036604166030884, | |
| "learning_rate": 7.427325581395349e-05, | |
| "loss": 0.0662, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 8.162790697674419, | |
| "grad_norm": 2.2601730823516846, | |
| "learning_rate": 7.412790697674419e-05, | |
| "loss": 0.0833, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 8.186046511627907, | |
| "grad_norm": 2.334266185760498, | |
| "learning_rate": 7.398255813953489e-05, | |
| "loss": 0.0482, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 8.209302325581396, | |
| "grad_norm": 5.374472141265869, | |
| "learning_rate": 7.383720930232558e-05, | |
| "loss": 0.1771, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 8.232558139534884, | |
| "grad_norm": 5.360229969024658, | |
| "learning_rate": 7.369186046511628e-05, | |
| "loss": 0.2253, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 8.255813953488373, | |
| "grad_norm": 2.7183334827423096, | |
| "learning_rate": 7.354651162790698e-05, | |
| "loss": 0.13, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 8.279069767441861, | |
| "grad_norm": 1.2531534433364868, | |
| "learning_rate": 7.340116279069768e-05, | |
| "loss": 0.0898, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 8.30232558139535, | |
| "grad_norm": 5.129657745361328, | |
| "learning_rate": 7.325581395348837e-05, | |
| "loss": 0.0884, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 8.325581395348838, | |
| "grad_norm": 0.5411895513534546, | |
| "learning_rate": 7.311046511627907e-05, | |
| "loss": 0.0227, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 8.348837209302326, | |
| "grad_norm": 3.1815500259399414, | |
| "learning_rate": 7.296511627906976e-05, | |
| "loss": 0.1153, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 8.372093023255815, | |
| "grad_norm": 3.956552743911743, | |
| "learning_rate": 7.281976744186046e-05, | |
| "loss": 0.1048, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 8.372093023255815, | |
| "eval_accuracy": 0.8918032786885246, | |
| "eval_f1": 0.8917846660000646, | |
| "eval_loss": 0.2492106705904007, | |
| "eval_precision": 0.8922580645161291, | |
| "eval_recall": 0.8928663737399845, | |
| "eval_runtime": 0.0761, | |
| "eval_samples_per_second": 4010.404, | |
| "eval_steps_per_second": 65.744, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 8.395348837209303, | |
| "grad_norm": 3.400632619857788, | |
| "learning_rate": 7.267441860465116e-05, | |
| "loss": 0.1008, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 8.418604651162791, | |
| "grad_norm": 4.204819202423096, | |
| "learning_rate": 7.252906976744186e-05, | |
| "loss": 0.0525, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 8.44186046511628, | |
| "grad_norm": 4.265756130218506, | |
| "learning_rate": 7.238372093023256e-05, | |
| "loss": 0.1789, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 8.465116279069768, | |
| "grad_norm": 3.7337512969970703, | |
| "learning_rate": 7.223837209302326e-05, | |
| "loss": 0.2187, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 8.488372093023255, | |
| "grad_norm": 7.506768226623535, | |
| "learning_rate": 7.209302325581396e-05, | |
| "loss": 0.1643, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 8.511627906976745, | |
| "grad_norm": 2.040639877319336, | |
| "learning_rate": 7.194767441860464e-05, | |
| "loss": 0.0645, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 8.534883720930232, | |
| "grad_norm": 3.809933662414551, | |
| "learning_rate": 7.180232558139535e-05, | |
| "loss": 0.1413, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 8.55813953488372, | |
| "grad_norm": 4.585285186767578, | |
| "learning_rate": 7.165697674418605e-05, | |
| "loss": 0.073, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 8.581395348837209, | |
| "grad_norm": 1.4243372678756714, | |
| "learning_rate": 7.151162790697675e-05, | |
| "loss": 0.0358, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 8.604651162790697, | |
| "grad_norm": 7.1119513511657715, | |
| "learning_rate": 7.136627906976745e-05, | |
| "loss": 0.1204, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 8.627906976744185, | |
| "grad_norm": 4.922991752624512, | |
| "learning_rate": 7.122093023255815e-05, | |
| "loss": 0.1183, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 8.651162790697674, | |
| "grad_norm": 2.3582427501678467, | |
| "learning_rate": 7.107558139534885e-05, | |
| "loss": 0.1057, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 8.674418604651162, | |
| "grad_norm": 1.6656956672668457, | |
| "learning_rate": 7.093023255813955e-05, | |
| "loss": 0.0434, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 8.69767441860465, | |
| "grad_norm": 3.274289608001709, | |
| "learning_rate": 7.078488372093023e-05, | |
| "loss": 0.1541, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 8.720930232558139, | |
| "grad_norm": 4.472874641418457, | |
| "learning_rate": 7.063953488372093e-05, | |
| "loss": 0.1322, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 8.744186046511627, | |
| "grad_norm": 2.455209493637085, | |
| "learning_rate": 7.049418604651163e-05, | |
| "loss": 0.1081, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 8.767441860465116, | |
| "grad_norm": 2.7206978797912598, | |
| "learning_rate": 7.034883720930233e-05, | |
| "loss": 0.1198, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 8.790697674418604, | |
| "grad_norm": 6.333116054534912, | |
| "learning_rate": 7.020348837209303e-05, | |
| "loss": 0.1965, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 8.813953488372093, | |
| "grad_norm": 1.0274301767349243, | |
| "learning_rate": 7.005813953488373e-05, | |
| "loss": 0.0388, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 8.837209302325581, | |
| "grad_norm": 4.29472541809082, | |
| "learning_rate": 6.991279069767443e-05, | |
| "loss": 0.0699, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 8.837209302325581, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8978290703580035, | |
| "eval_loss": 0.26147955656051636, | |
| "eval_precision": 0.9003302050747306, | |
| "eval_recall": 0.8969156543465151, | |
| "eval_runtime": 0.0748, | |
| "eval_samples_per_second": 4079.802, | |
| "eval_steps_per_second": 66.882, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 8.86046511627907, | |
| "grad_norm": 3.4688355922698975, | |
| "learning_rate": 6.976744186046513e-05, | |
| "loss": 0.1465, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 8.883720930232558, | |
| "grad_norm": 4.693603992462158, | |
| "learning_rate": 6.962209302325582e-05, | |
| "loss": 0.0675, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 8.906976744186046, | |
| "grad_norm": 1.98982572555542, | |
| "learning_rate": 6.947674418604651e-05, | |
| "loss": 0.1484, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 8.930232558139535, | |
| "grad_norm": 1.312723994255066, | |
| "learning_rate": 6.933139534883721e-05, | |
| "loss": 0.0458, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 8.953488372093023, | |
| "grad_norm": 2.106917142868042, | |
| "learning_rate": 6.918604651162791e-05, | |
| "loss": 0.0417, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 8.976744186046512, | |
| "grad_norm": 4.182035446166992, | |
| "learning_rate": 6.90406976744186e-05, | |
| "loss": 0.0459, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "grad_norm": 2.1285786628723145, | |
| "learning_rate": 6.88953488372093e-05, | |
| "loss": 0.0743, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 9.023255813953488, | |
| "grad_norm": 4.470812797546387, | |
| "learning_rate": 6.875e-05, | |
| "loss": 0.0973, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 9.046511627906977, | |
| "grad_norm": 1.6776498556137085, | |
| "learning_rate": 6.86046511627907e-05, | |
| "loss": 0.1894, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 9.069767441860465, | |
| "grad_norm": 3.9984216690063477, | |
| "learning_rate": 6.84593023255814e-05, | |
| "loss": 0.1704, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 9.093023255813954, | |
| "grad_norm": 2.3983538150787354, | |
| "learning_rate": 6.83139534883721e-05, | |
| "loss": 0.0736, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 9.116279069767442, | |
| "grad_norm": 2.989044427871704, | |
| "learning_rate": 6.81686046511628e-05, | |
| "loss": 0.0892, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 9.13953488372093, | |
| "grad_norm": 3.356992483139038, | |
| "learning_rate": 6.802325581395348e-05, | |
| "loss": 0.0999, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 9.162790697674419, | |
| "grad_norm": 2.2563319206237793, | |
| "learning_rate": 6.787790697674418e-05, | |
| "loss": 0.0452, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 9.186046511627907, | |
| "grad_norm": 5.215000152587891, | |
| "learning_rate": 6.773255813953488e-05, | |
| "loss": 0.0793, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 9.209302325581396, | |
| "grad_norm": 5.427519798278809, | |
| "learning_rate": 6.758720930232558e-05, | |
| "loss": 0.0755, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 9.232558139534884, | |
| "grad_norm": 7.729520797729492, | |
| "learning_rate": 6.744186046511628e-05, | |
| "loss": 0.2358, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 9.255813953488373, | |
| "grad_norm": 5.97319221496582, | |
| "learning_rate": 6.729651162790698e-05, | |
| "loss": 0.1829, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 9.279069767441861, | |
| "grad_norm": 4.662299633026123, | |
| "learning_rate": 6.715116279069768e-05, | |
| "loss": 0.0549, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 9.30232558139535, | |
| "grad_norm": 5.684605598449707, | |
| "learning_rate": 6.700581395348838e-05, | |
| "loss": 0.1466, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 9.30232558139535, | |
| "eval_accuracy": 0.9147540983606557, | |
| "eval_f1": 0.9145474137931033, | |
| "eval_loss": 0.22707484662532806, | |
| "eval_precision": 0.9148550724637681, | |
| "eval_recall": 0.9143189454639442, | |
| "eval_runtime": 0.0754, | |
| "eval_samples_per_second": 4044.013, | |
| "eval_steps_per_second": 66.295, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 9.325581395348838, | |
| "grad_norm": 8.901346206665039, | |
| "learning_rate": 6.686046511627908e-05, | |
| "loss": 0.1351, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 9.348837209302326, | |
| "grad_norm": 3.3197011947631836, | |
| "learning_rate": 6.671511627906976e-05, | |
| "loss": 0.0962, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 9.372093023255815, | |
| "grad_norm": 5.270369529724121, | |
| "learning_rate": 6.656976744186046e-05, | |
| "loss": 0.0862, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 9.395348837209303, | |
| "grad_norm": 1.4588712453842163, | |
| "learning_rate": 6.642441860465116e-05, | |
| "loss": 0.0286, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 9.418604651162791, | |
| "grad_norm": 7.928572177886963, | |
| "learning_rate": 6.627906976744186e-05, | |
| "loss": 0.1137, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 9.44186046511628, | |
| "grad_norm": 4.6285223960876465, | |
| "learning_rate": 6.613372093023256e-05, | |
| "loss": 0.0907, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 9.465116279069768, | |
| "grad_norm": 1.1047800779342651, | |
| "learning_rate": 6.598837209302326e-05, | |
| "loss": 0.0279, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 9.488372093023255, | |
| "grad_norm": 2.0423073768615723, | |
| "learning_rate": 6.584302325581395e-05, | |
| "loss": 0.0528, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 9.511627906976745, | |
| "grad_norm": 3.1448893547058105, | |
| "learning_rate": 6.569767441860465e-05, | |
| "loss": 0.0557, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 9.534883720930232, | |
| "grad_norm": 3.5887062549591064, | |
| "learning_rate": 6.555232558139535e-05, | |
| "loss": 0.0858, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 9.55813953488372, | |
| "grad_norm": 4.293898105621338, | |
| "learning_rate": 6.540697674418605e-05, | |
| "loss": 0.0547, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 9.581395348837209, | |
| "grad_norm": 3.3504300117492676, | |
| "learning_rate": 6.526162790697675e-05, | |
| "loss": 0.1361, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 9.604651162790697, | |
| "grad_norm": 0.5247292518615723, | |
| "learning_rate": 6.511627906976745e-05, | |
| "loss": 0.0168, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 9.627906976744185, | |
| "grad_norm": 0.9081286191940308, | |
| "learning_rate": 6.497093023255815e-05, | |
| "loss": 0.0337, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 9.651162790697674, | |
| "grad_norm": 5.076108932495117, | |
| "learning_rate": 6.482558139534885e-05, | |
| "loss": 0.0972, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 9.674418604651162, | |
| "grad_norm": 2.2447941303253174, | |
| "learning_rate": 6.468023255813955e-05, | |
| "loss": 0.0468, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 9.69767441860465, | |
| "grad_norm": 6.019301891326904, | |
| "learning_rate": 6.453488372093024e-05, | |
| "loss": 0.0967, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 9.720930232558139, | |
| "grad_norm": 1.4316606521606445, | |
| "learning_rate": 6.438953488372094e-05, | |
| "loss": 0.0666, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 9.744186046511627, | |
| "grad_norm": 1.9364207983016968, | |
| "learning_rate": 6.424418604651163e-05, | |
| "loss": 0.1047, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 9.767441860465116, | |
| "grad_norm": 4.00551700592041, | |
| "learning_rate": 6.409883720930233e-05, | |
| "loss": 0.0663, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 9.767441860465116, | |
| "eval_accuracy": 0.9114754098360656, | |
| "eval_f1": 0.9106105834464043, | |
| "eval_loss": 0.26964467763900757, | |
| "eval_precision": 0.9181201550387597, | |
| "eval_recall": 0.9089342638063238, | |
| "eval_runtime": 0.0726, | |
| "eval_samples_per_second": 4201.095, | |
| "eval_steps_per_second": 68.87, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 9.790697674418604, | |
| "grad_norm": 3.918790817260742, | |
| "learning_rate": 6.395348837209303e-05, | |
| "loss": 0.0773, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 9.813953488372093, | |
| "grad_norm": 4.5966901779174805, | |
| "learning_rate": 6.380813953488373e-05, | |
| "loss": 0.1099, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 9.837209302325581, | |
| "grad_norm": 3.332273244857788, | |
| "learning_rate": 6.366279069767442e-05, | |
| "loss": 0.0697, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 9.86046511627907, | |
| "grad_norm": 3.300297737121582, | |
| "learning_rate": 6.351744186046512e-05, | |
| "loss": 0.1235, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 9.883720930232558, | |
| "grad_norm": 1.9253029823303223, | |
| "learning_rate": 6.337209302325582e-05, | |
| "loss": 0.0276, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 9.906976744186046, | |
| "grad_norm": 2.729339599609375, | |
| "learning_rate": 6.322674418604652e-05, | |
| "loss": 0.2399, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 9.930232558139535, | |
| "grad_norm": 2.6790566444396973, | |
| "learning_rate": 6.308139534883722e-05, | |
| "loss": 0.0897, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 9.953488372093023, | |
| "grad_norm": 2.726901054382324, | |
| "learning_rate": 6.29360465116279e-05, | |
| "loss": 0.0768, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 9.976744186046512, | |
| "grad_norm": 2.4175329208374023, | |
| "learning_rate": 6.27906976744186e-05, | |
| "loss": 0.0943, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 4.533559799194336, | |
| "learning_rate": 6.26453488372093e-05, | |
| "loss": 0.127, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 10.023255813953488, | |
| "grad_norm": 3.8240880966186523, | |
| "learning_rate": 6.25e-05, | |
| "loss": 0.083, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 10.046511627906977, | |
| "grad_norm": 2.4785892963409424, | |
| "learning_rate": 6.23546511627907e-05, | |
| "loss": 0.0292, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 10.069767441860465, | |
| "grad_norm": 1.059528112411499, | |
| "learning_rate": 6.22093023255814e-05, | |
| "loss": 0.0619, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 10.093023255813954, | |
| "grad_norm": 2.8850722312927246, | |
| "learning_rate": 6.20639534883721e-05, | |
| "loss": 0.1071, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 10.116279069767442, | |
| "grad_norm": 1.9426606893539429, | |
| "learning_rate": 6.19186046511628e-05, | |
| "loss": 0.0988, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 10.13953488372093, | |
| "grad_norm": 2.1318485736846924, | |
| "learning_rate": 6.17732558139535e-05, | |
| "loss": 0.1346, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 10.162790697674419, | |
| "grad_norm": 2.793272018432617, | |
| "learning_rate": 6.162790697674418e-05, | |
| "loss": 0.0296, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 10.186046511627907, | |
| "grad_norm": 1.0038514137268066, | |
| "learning_rate": 6.148255813953488e-05, | |
| "loss": 0.0576, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 10.209302325581396, | |
| "grad_norm": 2.623504400253296, | |
| "learning_rate": 6.133720930232558e-05, | |
| "loss": 0.0607, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 10.232558139534884, | |
| "grad_norm": 1.1035139560699463, | |
| "learning_rate": 6.119186046511628e-05, | |
| "loss": 0.0473, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 10.232558139534884, | |
| "eval_accuracy": 0.9049180327868852, | |
| "eval_f1": 0.9045074224021592, | |
| "eval_loss": 0.2681082785129547, | |
| "eval_precision": 0.9062147872063795, | |
| "eval_recall": 0.9037649694150082, | |
| "eval_runtime": 0.0732, | |
| "eval_samples_per_second": 4164.7, | |
| "eval_steps_per_second": 68.274, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 10.255813953488373, | |
| "grad_norm": 4.833049774169922, | |
| "learning_rate": 6.104651162790698e-05, | |
| "loss": 0.0978, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 10.279069767441861, | |
| "grad_norm": 0.8670517802238464, | |
| "learning_rate": 6.0901162790697675e-05, | |
| "loss": 0.0284, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 10.30232558139535, | |
| "grad_norm": 4.661880016326904, | |
| "learning_rate": 6.0755813953488374e-05, | |
| "loss": 0.1335, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 10.325581395348838, | |
| "grad_norm": 4.113884449005127, | |
| "learning_rate": 6.0610465116279066e-05, | |
| "loss": 0.0624, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 10.348837209302326, | |
| "grad_norm": 4.014869689941406, | |
| "learning_rate": 6.0465116279069765e-05, | |
| "loss": 0.075, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 10.372093023255815, | |
| "grad_norm": 0.5572329759597778, | |
| "learning_rate": 6.0319767441860464e-05, | |
| "loss": 0.0172, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 10.395348837209303, | |
| "grad_norm": 2.2173776626586914, | |
| "learning_rate": 6.017441860465116e-05, | |
| "loss": 0.0753, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 10.418604651162791, | |
| "grad_norm": 4.567299842834473, | |
| "learning_rate": 6.002906976744186e-05, | |
| "loss": 0.1678, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 10.44186046511628, | |
| "grad_norm": 2.608336925506592, | |
| "learning_rate": 5.9883720930232554e-05, | |
| "loss": 0.1871, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 10.465116279069768, | |
| "grad_norm": 2.773268699645996, | |
| "learning_rate": 5.973837209302325e-05, | |
| "loss": 0.1034, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 10.488372093023255, | |
| "grad_norm": 4.1804304122924805, | |
| "learning_rate": 5.959302325581395e-05, | |
| "loss": 0.1966, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 10.511627906976745, | |
| "grad_norm": 2.119896650314331, | |
| "learning_rate": 5.944767441860465e-05, | |
| "loss": 0.0443, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 10.534883720930232, | |
| "grad_norm": 2.4072678089141846, | |
| "learning_rate": 5.9302325581395356e-05, | |
| "loss": 0.0509, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 10.55813953488372, | |
| "grad_norm": 2.580864906311035, | |
| "learning_rate": 5.9156976744186055e-05, | |
| "loss": 0.1404, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 10.581395348837209, | |
| "grad_norm": 1.998783826828003, | |
| "learning_rate": 5.9011627906976754e-05, | |
| "loss": 0.0546, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 10.604651162790697, | |
| "grad_norm": 1.9264705181121826, | |
| "learning_rate": 5.886627906976745e-05, | |
| "loss": 0.0753, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 10.627906976744185, | |
| "grad_norm": 1.8207919597625732, | |
| "learning_rate": 5.8720930232558145e-05, | |
| "loss": 0.0783, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 10.651162790697674, | |
| "grad_norm": 3.059663772583008, | |
| "learning_rate": 5.8575581395348844e-05, | |
| "loss": 0.0829, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 10.674418604651162, | |
| "grad_norm": 2.6100552082061768, | |
| "learning_rate": 5.843023255813954e-05, | |
| "loss": 0.1253, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 10.69767441860465, | |
| "grad_norm": 2.3080873489379883, | |
| "learning_rate": 5.828488372093024e-05, | |
| "loss": 0.0337, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 10.69767441860465, | |
| "eval_accuracy": 0.8918032786885246, | |
| "eval_f1": 0.8916355336398186, | |
| "eval_loss": 0.2402210682630539, | |
| "eval_precision": 0.8915439593558943, | |
| "eval_recall": 0.8917463599551995, | |
| "eval_runtime": 0.0746, | |
| "eval_samples_per_second": 4086.997, | |
| "eval_steps_per_second": 67.0, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 10.720930232558139, | |
| "grad_norm": 1.366176724433899, | |
| "learning_rate": 5.8139534883720933e-05, | |
| "loss": 0.0355, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 10.744186046511627, | |
| "grad_norm": 3.6105799674987793, | |
| "learning_rate": 5.799418604651163e-05, | |
| "loss": 0.1579, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 10.767441860465116, | |
| "grad_norm": 2.1533732414245605, | |
| "learning_rate": 5.784883720930233e-05, | |
| "loss": 0.0605, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 10.790697674418604, | |
| "grad_norm": 2.887242078781128, | |
| "learning_rate": 5.770348837209303e-05, | |
| "loss": 0.0787, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 10.813953488372093, | |
| "grad_norm": 4.048674583435059, | |
| "learning_rate": 5.755813953488373e-05, | |
| "loss": 0.0646, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 10.837209302325581, | |
| "grad_norm": 4.225156307220459, | |
| "learning_rate": 5.741279069767442e-05, | |
| "loss": 0.0449, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 10.86046511627907, | |
| "grad_norm": 5.0084123611450195, | |
| "learning_rate": 5.726744186046512e-05, | |
| "loss": 0.0798, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 10.883720930232558, | |
| "grad_norm": 2.2451436519622803, | |
| "learning_rate": 5.712209302325582e-05, | |
| "loss": 0.1, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 10.906976744186046, | |
| "grad_norm": 2.683537006378174, | |
| "learning_rate": 5.697674418604652e-05, | |
| "loss": 0.0875, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 10.930232558139535, | |
| "grad_norm": 2.50972580909729, | |
| "learning_rate": 5.683139534883721e-05, | |
| "loss": 0.0313, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 10.953488372093023, | |
| "grad_norm": 8.650345802307129, | |
| "learning_rate": 5.668604651162791e-05, | |
| "loss": 0.0685, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 10.976744186046512, | |
| "grad_norm": 2.9768123626708984, | |
| "learning_rate": 5.654069767441861e-05, | |
| "loss": 0.031, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "grad_norm": 2.2081665992736816, | |
| "learning_rate": 5.6395348837209306e-05, | |
| "loss": 0.0335, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 11.023255813953488, | |
| "grad_norm": 4.979495048522949, | |
| "learning_rate": 5.6250000000000005e-05, | |
| "loss": 0.0331, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 11.046511627906977, | |
| "grad_norm": 2.000579833984375, | |
| "learning_rate": 5.61046511627907e-05, | |
| "loss": 0.0472, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 11.069767441860465, | |
| "grad_norm": 2.3896875381469727, | |
| "learning_rate": 5.5959302325581396e-05, | |
| "loss": 0.0745, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 11.093023255813954, | |
| "grad_norm": 2.0320870876312256, | |
| "learning_rate": 5.5813953488372095e-05, | |
| "loss": 0.0634, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 11.116279069767442, | |
| "grad_norm": 2.134622812271118, | |
| "learning_rate": 5.5668604651162794e-05, | |
| "loss": 0.0387, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 11.13953488372093, | |
| "grad_norm": 1.0899347066879272, | |
| "learning_rate": 5.552325581395349e-05, | |
| "loss": 0.0455, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 11.162790697674419, | |
| "grad_norm": 0.905604898929596, | |
| "learning_rate": 5.5377906976744185e-05, | |
| "loss": 0.0156, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 11.162790697674419, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9008624425760596, | |
| "eval_loss": 0.3055068254470825, | |
| "eval_precision": 0.906113154668068, | |
| "eval_recall": 0.8995003015421728, | |
| "eval_runtime": 0.0798, | |
| "eval_samples_per_second": 3820.2, | |
| "eval_steps_per_second": 62.626, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 11.186046511627907, | |
| "grad_norm": 1.7722190618515015, | |
| "learning_rate": 5.5232558139534884e-05, | |
| "loss": 0.0672, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 11.209302325581396, | |
| "grad_norm": 3.3152947425842285, | |
| "learning_rate": 5.508720930232558e-05, | |
| "loss": 0.1093, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 11.232558139534884, | |
| "grad_norm": 3.6544511318206787, | |
| "learning_rate": 5.494186046511628e-05, | |
| "loss": 0.0894, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 11.255813953488373, | |
| "grad_norm": 4.428676128387451, | |
| "learning_rate": 5.4796511627906974e-05, | |
| "loss": 0.0495, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 11.279069767441861, | |
| "grad_norm": 3.5235519409179688, | |
| "learning_rate": 5.465116279069767e-05, | |
| "loss": 0.0313, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 11.30232558139535, | |
| "grad_norm": 1.8089061975479126, | |
| "learning_rate": 5.450581395348837e-05, | |
| "loss": 0.0294, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 11.325581395348838, | |
| "grad_norm": 9.173127174377441, | |
| "learning_rate": 5.436046511627907e-05, | |
| "loss": 0.0559, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 11.348837209302326, | |
| "grad_norm": 10.084165573120117, | |
| "learning_rate": 5.421511627906977e-05, | |
| "loss": 0.0759, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 11.372093023255815, | |
| "grad_norm": 0.5530898571014404, | |
| "learning_rate": 5.406976744186046e-05, | |
| "loss": 0.0202, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 11.395348837209303, | |
| "grad_norm": 4.215567588806152, | |
| "learning_rate": 5.392441860465116e-05, | |
| "loss": 0.1209, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 11.418604651162791, | |
| "grad_norm": 0.7583315372467041, | |
| "learning_rate": 5.377906976744186e-05, | |
| "loss": 0.0054, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 11.44186046511628, | |
| "grad_norm": 0.2435784786939621, | |
| "learning_rate": 5.363372093023256e-05, | |
| "loss": 0.0064, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 11.465116279069768, | |
| "grad_norm": 2.534301519393921, | |
| "learning_rate": 5.348837209302326e-05, | |
| "loss": 0.0733, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 11.488372093023255, | |
| "grad_norm": 3.94118332862854, | |
| "learning_rate": 5.334302325581395e-05, | |
| "loss": 0.0812, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 11.511627906976745, | |
| "grad_norm": 3.799800157546997, | |
| "learning_rate": 5.319767441860465e-05, | |
| "loss": 0.1377, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 11.534883720930232, | |
| "grad_norm": 0.16743521392345428, | |
| "learning_rate": 5.305232558139536e-05, | |
| "loss": 0.0054, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 11.55813953488372, | |
| "grad_norm": 1.3862305879592896, | |
| "learning_rate": 5.290697674418605e-05, | |
| "loss": 0.0364, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 11.581395348837209, | |
| "grad_norm": 3.8730082511901855, | |
| "learning_rate": 5.276162790697675e-05, | |
| "loss": 0.0827, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 11.604651162790697, | |
| "grad_norm": 5.6961188316345215, | |
| "learning_rate": 5.261627906976745e-05, | |
| "loss": 0.0936, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 11.627906976744185, | |
| "grad_norm": 0.9591822028160095, | |
| "learning_rate": 5.247093023255815e-05, | |
| "loss": 0.0146, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 11.627906976744185, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.900974025974026, | |
| "eval_loss": 0.32393914461135864, | |
| "eval_precision": 0.9050144016758314, | |
| "eval_recall": 0.899780304988369, | |
| "eval_runtime": 0.075, | |
| "eval_samples_per_second": 4066.651, | |
| "eval_steps_per_second": 66.666, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 11.651162790697674, | |
| "grad_norm": 5.92799186706543, | |
| "learning_rate": 5.232558139534884e-05, | |
| "loss": 0.0904, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 11.674418604651162, | |
| "grad_norm": 9.12846851348877, | |
| "learning_rate": 5.218023255813954e-05, | |
| "loss": 0.322, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 11.69767441860465, | |
| "grad_norm": 2.95119571685791, | |
| "learning_rate": 5.203488372093024e-05, | |
| "loss": 0.0175, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 11.720930232558139, | |
| "grad_norm": 3.9121649265289307, | |
| "learning_rate": 5.188953488372094e-05, | |
| "loss": 0.0733, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 11.744186046511627, | |
| "grad_norm": 2.7404439449310303, | |
| "learning_rate": 5.1744186046511636e-05, | |
| "loss": 0.0178, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 11.767441860465116, | |
| "grad_norm": 4.207398891448975, | |
| "learning_rate": 5.159883720930233e-05, | |
| "loss": 0.1179, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 11.790697674418604, | |
| "grad_norm": 5.63287878036499, | |
| "learning_rate": 5.145348837209303e-05, | |
| "loss": 0.0728, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 11.813953488372093, | |
| "grad_norm": 4.2650556564331055, | |
| "learning_rate": 5.1308139534883726e-05, | |
| "loss": 0.0409, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 11.837209302325581, | |
| "grad_norm": 4.405393123626709, | |
| "learning_rate": 5.1162790697674425e-05, | |
| "loss": 0.1272, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 11.86046511627907, | |
| "grad_norm": 3.2369868755340576, | |
| "learning_rate": 5.1017441860465124e-05, | |
| "loss": 0.1354, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 11.883720930232558, | |
| "grad_norm": 2.492950439453125, | |
| "learning_rate": 5.0872093023255816e-05, | |
| "loss": 0.116, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 11.906976744186046, | |
| "grad_norm": 5.160562992095947, | |
| "learning_rate": 5.0726744186046515e-05, | |
| "loss": 0.0813, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 11.930232558139535, | |
| "grad_norm": 3.234938383102417, | |
| "learning_rate": 5.0581395348837214e-05, | |
| "loss": 0.041, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 11.953488372093023, | |
| "grad_norm": 2.3901174068450928, | |
| "learning_rate": 5.043604651162791e-05, | |
| "loss": 0.0285, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 11.976744186046512, | |
| "grad_norm": 1.9587090015411377, | |
| "learning_rate": 5.0290697674418605e-05, | |
| "loss": 0.0191, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "grad_norm": 1.5882123708724976, | |
| "learning_rate": 5.0145348837209304e-05, | |
| "loss": 0.0292, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 12.023255813953488, | |
| "grad_norm": 1.0756316184997559, | |
| "learning_rate": 5e-05, | |
| "loss": 0.0348, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 12.046511627906977, | |
| "grad_norm": 7.145483016967773, | |
| "learning_rate": 4.98546511627907e-05, | |
| "loss": 0.1204, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 12.069767441860465, | |
| "grad_norm": 10.459267616271973, | |
| "learning_rate": 4.97093023255814e-05, | |
| "loss": 0.0682, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 12.093023255813954, | |
| "grad_norm": 2.6111533641815186, | |
| "learning_rate": 4.956395348837209e-05, | |
| "loss": 0.1088, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 12.093023255813954, | |
| "eval_accuracy": 0.9081967213114754, | |
| "eval_f1": 0.907471613070989, | |
| "eval_loss": 0.29710283875465393, | |
| "eval_precision": 0.9127912068663514, | |
| "eval_recall": 0.9060696131644697, | |
| "eval_runtime": 0.0726, | |
| "eval_samples_per_second": 4200.336, | |
| "eval_steps_per_second": 68.858, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 12.116279069767442, | |
| "grad_norm": 1.2339119911193848, | |
| "learning_rate": 4.941860465116279e-05, | |
| "loss": 0.0222, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 12.13953488372093, | |
| "grad_norm": 1.6148200035095215, | |
| "learning_rate": 4.927325581395349e-05, | |
| "loss": 0.0968, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 12.162790697674419, | |
| "grad_norm": 2.2320408821105957, | |
| "learning_rate": 4.912790697674419e-05, | |
| "loss": 0.0742, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 12.186046511627907, | |
| "grad_norm": 1.0352402925491333, | |
| "learning_rate": 4.898255813953488e-05, | |
| "loss": 0.0287, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 12.209302325581396, | |
| "grad_norm": 3.2948336601257324, | |
| "learning_rate": 4.883720930232558e-05, | |
| "loss": 0.0945, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 12.232558139534884, | |
| "grad_norm": 0.48531392216682434, | |
| "learning_rate": 4.869186046511628e-05, | |
| "loss": 0.0119, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 12.255813953488373, | |
| "grad_norm": 3.030161142349243, | |
| "learning_rate": 4.854651162790698e-05, | |
| "loss": 0.0713, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 12.279069767441861, | |
| "grad_norm": 1.6946715116500854, | |
| "learning_rate": 4.8401162790697676e-05, | |
| "loss": 0.0322, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 12.30232558139535, | |
| "grad_norm": 4.674280643463135, | |
| "learning_rate": 4.8255813953488375e-05, | |
| "loss": 0.048, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 12.325581395348838, | |
| "grad_norm": 1.2208576202392578, | |
| "learning_rate": 4.8110465116279074e-05, | |
| "loss": 0.009, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 12.348837209302326, | |
| "grad_norm": 7.181972026824951, | |
| "learning_rate": 4.796511627906977e-05, | |
| "loss": 0.0275, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 12.372093023255815, | |
| "grad_norm": 1.5175693035125732, | |
| "learning_rate": 4.781976744186047e-05, | |
| "loss": 0.0246, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 12.395348837209303, | |
| "grad_norm": 5.098257064819336, | |
| "learning_rate": 4.7674418604651164e-05, | |
| "loss": 0.0758, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 12.418604651162791, | |
| "grad_norm": 2.7552084922790527, | |
| "learning_rate": 4.752906976744186e-05, | |
| "loss": 0.0558, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 12.44186046511628, | |
| "grad_norm": 2.456735134124756, | |
| "learning_rate": 4.738372093023256e-05, | |
| "loss": 0.0484, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 12.465116279069768, | |
| "grad_norm": 1.6237318515777588, | |
| "learning_rate": 4.723837209302326e-05, | |
| "loss": 0.08, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 12.488372093023255, | |
| "grad_norm": 3.8842220306396484, | |
| "learning_rate": 4.709302325581396e-05, | |
| "loss": 0.0858, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 12.511627906976745, | |
| "grad_norm": 3.0373013019561768, | |
| "learning_rate": 4.694767441860465e-05, | |
| "loss": 0.0256, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 12.534883720930232, | |
| "grad_norm": 5.356001853942871, | |
| "learning_rate": 4.680232558139535e-05, | |
| "loss": 0.1057, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 12.55813953488372, | |
| "grad_norm": 0.7480242252349854, | |
| "learning_rate": 4.665697674418605e-05, | |
| "loss": 0.0102, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 12.55813953488372, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8983213068212369, | |
| "eval_loss": 0.27956801652908325, | |
| "eval_precision": 0.8984348125214998, | |
| "eval_recall": 0.8991556819160851, | |
| "eval_runtime": 0.0733, | |
| "eval_samples_per_second": 4158.554, | |
| "eval_steps_per_second": 68.173, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 12.581395348837209, | |
| "grad_norm": 8.634737014770508, | |
| "learning_rate": 4.651162790697675e-05, | |
| "loss": 0.1041, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 12.604651162790697, | |
| "grad_norm": 2.154197931289673, | |
| "learning_rate": 4.636627906976744e-05, | |
| "loss": 0.0883, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 12.627906976744185, | |
| "grad_norm": 5.6526265144348145, | |
| "learning_rate": 4.622093023255814e-05, | |
| "loss": 0.0634, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 12.651162790697674, | |
| "grad_norm": 0.4163094162940979, | |
| "learning_rate": 4.607558139534884e-05, | |
| "loss": 0.0102, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 12.674418604651162, | |
| "grad_norm": 2.6310293674468994, | |
| "learning_rate": 4.593023255813954e-05, | |
| "loss": 0.0613, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 12.69767441860465, | |
| "grad_norm": 0.8978974223136902, | |
| "learning_rate": 4.5784883720930236e-05, | |
| "loss": 0.0118, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 12.720930232558139, | |
| "grad_norm": 3.0897459983825684, | |
| "learning_rate": 4.563953488372093e-05, | |
| "loss": 0.0592, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 12.744186046511627, | |
| "grad_norm": 1.5553648471832275, | |
| "learning_rate": 4.549418604651163e-05, | |
| "loss": 0.0357, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 12.767441860465116, | |
| "grad_norm": 1.9957389831542969, | |
| "learning_rate": 4.5348837209302326e-05, | |
| "loss": 0.0415, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 12.790697674418604, | |
| "grad_norm": 0.22972577810287476, | |
| "learning_rate": 4.520348837209303e-05, | |
| "loss": 0.0078, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 12.813953488372093, | |
| "grad_norm": 2.78206205368042, | |
| "learning_rate": 4.505813953488372e-05, | |
| "loss": 0.026, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 12.837209302325581, | |
| "grad_norm": 3.2643375396728516, | |
| "learning_rate": 4.491279069767442e-05, | |
| "loss": 0.0386, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 12.86046511627907, | |
| "grad_norm": 9.664105415344238, | |
| "learning_rate": 4.476744186046512e-05, | |
| "loss": 0.1431, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 12.883720930232558, | |
| "grad_norm": 3.2015540599823, | |
| "learning_rate": 4.462209302325582e-05, | |
| "loss": 0.0605, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 12.906976744186046, | |
| "grad_norm": 4.17910099029541, | |
| "learning_rate": 4.447674418604651e-05, | |
| "loss": 0.0797, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 12.930232558139535, | |
| "grad_norm": 3.999985933303833, | |
| "learning_rate": 4.433139534883721e-05, | |
| "loss": 0.117, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 12.953488372093023, | |
| "grad_norm": 1.1732484102249146, | |
| "learning_rate": 4.418604651162791e-05, | |
| "loss": 0.0244, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 12.976744186046512, | |
| "grad_norm": 4.184539794921875, | |
| "learning_rate": 4.404069767441861e-05, | |
| "loss": 0.0517, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "grad_norm": 2.3533170223236084, | |
| "learning_rate": 4.389534883720931e-05, | |
| "loss": 0.0505, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 13.023255813953488, | |
| "grad_norm": 2.3821420669555664, | |
| "learning_rate": 4.375e-05, | |
| "loss": 0.0253, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 13.023255813953488, | |
| "eval_accuracy": 0.8885245901639345, | |
| "eval_f1": 0.8879936058066189, | |
| "eval_loss": 0.3079434633255005, | |
| "eval_precision": 0.8900242992276317, | |
| "eval_recall": 0.8872016886361678, | |
| "eval_runtime": 0.0743, | |
| "eval_samples_per_second": 4102.318, | |
| "eval_steps_per_second": 67.251, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 13.046511627906977, | |
| "grad_norm": 3.6151607036590576, | |
| "learning_rate": 4.36046511627907e-05, | |
| "loss": 0.08, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 13.069767441860465, | |
| "grad_norm": 1.870180368423462, | |
| "learning_rate": 4.34593023255814e-05, | |
| "loss": 0.0475, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 13.093023255813954, | |
| "grad_norm": 3.9536831378936768, | |
| "learning_rate": 4.3313953488372096e-05, | |
| "loss": 0.0206, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 13.116279069767442, | |
| "grad_norm": 6.368700981140137, | |
| "learning_rate": 4.3168604651162795e-05, | |
| "loss": 0.0485, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 13.13953488372093, | |
| "grad_norm": 1.1507411003112793, | |
| "learning_rate": 4.302325581395349e-05, | |
| "loss": 0.0133, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 13.162790697674419, | |
| "grad_norm": 1.034654140472412, | |
| "learning_rate": 4.2877906976744186e-05, | |
| "loss": 0.0437, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 13.186046511627907, | |
| "grad_norm": 0.2083161622285843, | |
| "learning_rate": 4.2732558139534885e-05, | |
| "loss": 0.0054, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 13.209302325581396, | |
| "grad_norm": 1.1866053342819214, | |
| "learning_rate": 4.2587209302325584e-05, | |
| "loss": 0.0134, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 13.232558139534884, | |
| "grad_norm": 0.6963038444519043, | |
| "learning_rate": 4.2441860465116276e-05, | |
| "loss": 0.0083, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 13.255813953488373, | |
| "grad_norm": 3.132704973220825, | |
| "learning_rate": 4.2296511627906975e-05, | |
| "loss": 0.0567, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 13.279069767441861, | |
| "grad_norm": 1.62773859500885, | |
| "learning_rate": 4.215116279069768e-05, | |
| "loss": 0.0803, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 13.30232558139535, | |
| "grad_norm": 0.7541981935501099, | |
| "learning_rate": 4.200581395348838e-05, | |
| "loss": 0.0114, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 13.325581395348838, | |
| "grad_norm": 5.177672386169434, | |
| "learning_rate": 4.186046511627907e-05, | |
| "loss": 0.0302, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 13.348837209302326, | |
| "grad_norm": 3.653960704803467, | |
| "learning_rate": 4.171511627906977e-05, | |
| "loss": 0.0648, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 13.372093023255815, | |
| "grad_norm": 0.6348634362220764, | |
| "learning_rate": 4.156976744186047e-05, | |
| "loss": 0.0077, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 13.395348837209303, | |
| "grad_norm": 4.980737686157227, | |
| "learning_rate": 4.142441860465117e-05, | |
| "loss": 0.0857, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 13.418604651162791, | |
| "grad_norm": 4.5483551025390625, | |
| "learning_rate": 4.127906976744187e-05, | |
| "loss": 0.0456, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 13.44186046511628, | |
| "grad_norm": 4.054844379425049, | |
| "learning_rate": 4.113372093023256e-05, | |
| "loss": 0.0837, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 13.465116279069768, | |
| "grad_norm": 3.3397719860076904, | |
| "learning_rate": 4.098837209302326e-05, | |
| "loss": 0.0742, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 13.488372093023255, | |
| "grad_norm": 6.787147045135498, | |
| "learning_rate": 4.0843023255813957e-05, | |
| "loss": 0.0868, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 13.488372093023255, | |
| "eval_accuracy": 0.8918032786885246, | |
| "eval_f1": 0.8914251194701244, | |
| "eval_loss": 0.34987983107566833, | |
| "eval_precision": 0.8924277806607853, | |
| "eval_recall": 0.8909063496166107, | |
| "eval_runtime": 0.0747, | |
| "eval_samples_per_second": 4085.209, | |
| "eval_steps_per_second": 66.971, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 13.511627906976745, | |
| "grad_norm": 1.7752641439437866, | |
| "learning_rate": 4.0697674418604655e-05, | |
| "loss": 0.0476, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 13.534883720930232, | |
| "grad_norm": 2.80531907081604, | |
| "learning_rate": 4.055232558139535e-05, | |
| "loss": 0.0202, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 13.55813953488372, | |
| "grad_norm": 0.5729110836982727, | |
| "learning_rate": 4.0406976744186046e-05, | |
| "loss": 0.0078, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 13.581395348837209, | |
| "grad_norm": 5.221317768096924, | |
| "learning_rate": 4.0261627906976745e-05, | |
| "loss": 0.092, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 13.604651162790697, | |
| "grad_norm": 3.757556915283203, | |
| "learning_rate": 4.0116279069767444e-05, | |
| "loss": 0.024, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 13.627906976744185, | |
| "grad_norm": 6.441154479980469, | |
| "learning_rate": 3.997093023255814e-05, | |
| "loss": 0.1422, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 13.651162790697674, | |
| "grad_norm": 2.314418077468872, | |
| "learning_rate": 3.9825581395348835e-05, | |
| "loss": 0.0815, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 13.674418604651162, | |
| "grad_norm": 2.129340887069702, | |
| "learning_rate": 3.9680232558139534e-05, | |
| "loss": 0.0419, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 13.69767441860465, | |
| "grad_norm": 1.477169394493103, | |
| "learning_rate": 3.953488372093023e-05, | |
| "loss": 0.0379, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 13.720930232558139, | |
| "grad_norm": 2.567399740219116, | |
| "learning_rate": 3.938953488372093e-05, | |
| "loss": 0.1169, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 13.744186046511627, | |
| "grad_norm": 1.2707808017730713, | |
| "learning_rate": 3.924418604651163e-05, | |
| "loss": 0.0225, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 13.767441860465116, | |
| "grad_norm": 0.3642787039279938, | |
| "learning_rate": 3.909883720930232e-05, | |
| "loss": 0.0053, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 13.790697674418604, | |
| "grad_norm": 2.2786664962768555, | |
| "learning_rate": 3.895348837209303e-05, | |
| "loss": 0.043, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 13.813953488372093, | |
| "grad_norm": 6.119050979614258, | |
| "learning_rate": 3.880813953488373e-05, | |
| "loss": 0.1129, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 13.837209302325581, | |
| "grad_norm": 3.127246141433716, | |
| "learning_rate": 3.866279069767442e-05, | |
| "loss": 0.1094, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 13.86046511627907, | |
| "grad_norm": 4.50978946685791, | |
| "learning_rate": 3.851744186046512e-05, | |
| "loss": 0.0714, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 13.883720930232558, | |
| "grad_norm": 2.3526670932769775, | |
| "learning_rate": 3.837209302325582e-05, | |
| "loss": 0.1126, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 13.906976744186046, | |
| "grad_norm": 4.066991806030273, | |
| "learning_rate": 3.8226744186046516e-05, | |
| "loss": 0.0466, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 13.930232558139535, | |
| "grad_norm": 1.5120619535446167, | |
| "learning_rate": 3.8081395348837215e-05, | |
| "loss": 0.0609, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 13.953488372093023, | |
| "grad_norm": 3.2349016666412354, | |
| "learning_rate": 3.793604651162791e-05, | |
| "loss": 0.0399, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 13.953488372093023, | |
| "eval_accuracy": 0.8885245901639345, | |
| "eval_f1": 0.8880903047569713, | |
| "eval_loss": 0.36928993463516235, | |
| "eval_precision": 0.8893939393939394, | |
| "eval_recall": 0.887481692082364, | |
| "eval_runtime": 0.0753, | |
| "eval_samples_per_second": 4048.646, | |
| "eval_steps_per_second": 66.371, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 13.976744186046512, | |
| "grad_norm": 2.7635951042175293, | |
| "learning_rate": 3.7790697674418606e-05, | |
| "loss": 0.1149, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "grad_norm": 0.716441810131073, | |
| "learning_rate": 3.7645348837209305e-05, | |
| "loss": 0.0095, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 14.023255813953488, | |
| "grad_norm": 1.099993348121643, | |
| "learning_rate": 3.7500000000000003e-05, | |
| "loss": 0.0179, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 14.046511627906977, | |
| "grad_norm": 0.9968247413635254, | |
| "learning_rate": 3.73546511627907e-05, | |
| "loss": 0.0295, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 14.069767441860465, | |
| "grad_norm": 0.9443352222442627, | |
| "learning_rate": 3.7209302325581394e-05, | |
| "loss": 0.0121, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 14.093023255813954, | |
| "grad_norm": 0.24217644333839417, | |
| "learning_rate": 3.706395348837209e-05, | |
| "loss": 0.0066, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 14.116279069767442, | |
| "grad_norm": 0.32041826844215393, | |
| "learning_rate": 3.691860465116279e-05, | |
| "loss": 0.0071, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 14.13953488372093, | |
| "grad_norm": 4.715015411376953, | |
| "learning_rate": 3.677325581395349e-05, | |
| "loss": 0.0775, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 14.162790697674419, | |
| "grad_norm": 1.9597333669662476, | |
| "learning_rate": 3.662790697674418e-05, | |
| "loss": 0.0232, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 14.186046511627907, | |
| "grad_norm": 1.3680325746536255, | |
| "learning_rate": 3.648255813953488e-05, | |
| "loss": 0.0194, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 14.209302325581396, | |
| "grad_norm": 7.5977396965026855, | |
| "learning_rate": 3.633720930232558e-05, | |
| "loss": 0.1274, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 14.232558139534884, | |
| "grad_norm": 7.637692451477051, | |
| "learning_rate": 3.619186046511628e-05, | |
| "loss": 0.1341, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 14.255813953488373, | |
| "grad_norm": 10.332743644714355, | |
| "learning_rate": 3.604651162790698e-05, | |
| "loss": 0.1039, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 14.279069767441861, | |
| "grad_norm": 1.6394683122634888, | |
| "learning_rate": 3.590116279069768e-05, | |
| "loss": 0.0126, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 14.30232558139535, | |
| "grad_norm": 1.69204580783844, | |
| "learning_rate": 3.5755813953488376e-05, | |
| "loss": 0.0327, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 14.325581395348838, | |
| "grad_norm": 3.1536524295806885, | |
| "learning_rate": 3.5610465116279075e-05, | |
| "loss": 0.0347, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 14.348837209302326, | |
| "grad_norm": 0.6318650841712952, | |
| "learning_rate": 3.5465116279069774e-05, | |
| "loss": 0.0424, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 14.372093023255815, | |
| "grad_norm": 5.714948654174805, | |
| "learning_rate": 3.5319767441860466e-05, | |
| "loss": 0.0339, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 14.395348837209303, | |
| "grad_norm": 4.141314506530762, | |
| "learning_rate": 3.5174418604651165e-05, | |
| "loss": 0.045, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 14.418604651162791, | |
| "grad_norm": 1.1429768800735474, | |
| "learning_rate": 3.5029069767441864e-05, | |
| "loss": 0.0291, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 14.418604651162791, | |
| "eval_accuracy": 0.8950819672131147, | |
| "eval_f1": 0.8944819235426397, | |
| "eval_loss": 0.3720164895057678, | |
| "eval_precision": 0.8974286460146188, | |
| "eval_recall": 0.8934909968122684, | |
| "eval_runtime": 0.0902, | |
| "eval_samples_per_second": 3381.6, | |
| "eval_steps_per_second": 55.436, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 14.44186046511628, | |
| "grad_norm": 11.266057014465332, | |
| "learning_rate": 3.488372093023256e-05, | |
| "loss": 0.0526, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 14.465116279069768, | |
| "grad_norm": 0.8715966939926147, | |
| "learning_rate": 3.4738372093023255e-05, | |
| "loss": 0.0256, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 14.488372093023255, | |
| "grad_norm": 3.1204757690429688, | |
| "learning_rate": 3.4593023255813954e-05, | |
| "loss": 0.0256, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 14.511627906976745, | |
| "grad_norm": 2.438506841659546, | |
| "learning_rate": 3.444767441860465e-05, | |
| "loss": 0.0321, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 14.534883720930232, | |
| "grad_norm": 3.6766304969787598, | |
| "learning_rate": 3.430232558139535e-05, | |
| "loss": 0.1027, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 14.55813953488372, | |
| "grad_norm": 6.79380464553833, | |
| "learning_rate": 3.415697674418605e-05, | |
| "loss": 0.0597, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 14.581395348837209, | |
| "grad_norm": 1.6378151178359985, | |
| "learning_rate": 3.401162790697674e-05, | |
| "loss": 0.0238, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 14.604651162790697, | |
| "grad_norm": 0.5451012253761292, | |
| "learning_rate": 3.386627906976744e-05, | |
| "loss": 0.0103, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 14.627906976744185, | |
| "grad_norm": 2.0606513023376465, | |
| "learning_rate": 3.372093023255814e-05, | |
| "loss": 0.0223, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 14.651162790697674, | |
| "grad_norm": 0.4602338373661041, | |
| "learning_rate": 3.357558139534884e-05, | |
| "loss": 0.0057, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 14.674418604651162, | |
| "grad_norm": 4.629768371582031, | |
| "learning_rate": 3.343023255813954e-05, | |
| "loss": 0.1183, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 14.69767441860465, | |
| "grad_norm": 6.286869525909424, | |
| "learning_rate": 3.328488372093023e-05, | |
| "loss": 0.0308, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 14.720930232558139, | |
| "grad_norm": 1.118071436882019, | |
| "learning_rate": 3.313953488372093e-05, | |
| "loss": 0.0692, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 14.744186046511627, | |
| "grad_norm": 2.621478319168091, | |
| "learning_rate": 3.299418604651163e-05, | |
| "loss": 0.0206, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 14.767441860465116, | |
| "grad_norm": 2.408926486968994, | |
| "learning_rate": 3.284883720930233e-05, | |
| "loss": 0.0868, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 14.790697674418604, | |
| "grad_norm": 4.2226104736328125, | |
| "learning_rate": 3.2703488372093026e-05, | |
| "loss": 0.0165, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 14.813953488372093, | |
| "grad_norm": 8.624298095703125, | |
| "learning_rate": 3.2558139534883724e-05, | |
| "loss": 0.1733, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 14.837209302325581, | |
| "grad_norm": 5.992215633392334, | |
| "learning_rate": 3.241279069767442e-05, | |
| "loss": 0.0286, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 14.86046511627907, | |
| "grad_norm": 0.8743571043014526, | |
| "learning_rate": 3.226744186046512e-05, | |
| "loss": 0.0113, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 14.883720930232558, | |
| "grad_norm": 0.9226441979408264, | |
| "learning_rate": 3.2122093023255814e-05, | |
| "loss": 0.0069, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 14.883720930232558, | |
| "eval_accuracy": 0.8918032786885246, | |
| "eval_f1": 0.8915047052355851, | |
| "eval_loss": 0.3578951358795166, | |
| "eval_precision": 0.8919968919968919, | |
| "eval_recall": 0.8911863530628069, | |
| "eval_runtime": 0.0756, | |
| "eval_samples_per_second": 4035.886, | |
| "eval_steps_per_second": 66.162, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 14.906976744186046, | |
| "grad_norm": 4.145755767822266, | |
| "learning_rate": 3.197674418604651e-05, | |
| "loss": 0.0439, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 14.930232558139535, | |
| "grad_norm": 1.2993311882019043, | |
| "learning_rate": 3.183139534883721e-05, | |
| "loss": 0.0446, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 14.953488372093023, | |
| "grad_norm": 0.2827296555042267, | |
| "learning_rate": 3.168604651162791e-05, | |
| "loss": 0.0052, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 14.976744186046512, | |
| "grad_norm": 2.9335291385650635, | |
| "learning_rate": 3.154069767441861e-05, | |
| "loss": 0.0644, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 7.6974968910217285, | |
| "learning_rate": 3.13953488372093e-05, | |
| "loss": 0.0456, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 15.023255813953488, | |
| "grad_norm": 2.137190103530884, | |
| "learning_rate": 3.125e-05, | |
| "loss": 0.0147, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 15.046511627906977, | |
| "grad_norm": 0.834999680519104, | |
| "learning_rate": 3.11046511627907e-05, | |
| "loss": 0.0181, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 15.069767441860465, | |
| "grad_norm": 0.8868411183357239, | |
| "learning_rate": 3.09593023255814e-05, | |
| "loss": 0.0467, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 15.093023255813954, | |
| "grad_norm": 3.1377875804901123, | |
| "learning_rate": 3.081395348837209e-05, | |
| "loss": 0.0281, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 15.116279069767442, | |
| "grad_norm": 3.3987367153167725, | |
| "learning_rate": 3.066860465116279e-05, | |
| "loss": 0.1133, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 15.13953488372093, | |
| "grad_norm": 0.19906505942344666, | |
| "learning_rate": 3.052325581395349e-05, | |
| "loss": 0.0032, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 15.162790697674419, | |
| "grad_norm": 4.011625289916992, | |
| "learning_rate": 3.0377906976744187e-05, | |
| "loss": 0.0792, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 15.186046511627907, | |
| "grad_norm": 9.421031951904297, | |
| "learning_rate": 3.0232558139534883e-05, | |
| "loss": 0.0612, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 15.209302325581396, | |
| "grad_norm": 3.5618064403533936, | |
| "learning_rate": 3.008720930232558e-05, | |
| "loss": 0.0159, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 15.232558139534884, | |
| "grad_norm": 7.569765090942383, | |
| "learning_rate": 2.9941860465116277e-05, | |
| "loss": 0.1039, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 15.255813953488373, | |
| "grad_norm": 6.453846454620361, | |
| "learning_rate": 2.9796511627906976e-05, | |
| "loss": 0.0283, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 15.279069767441861, | |
| "grad_norm": 0.1914948970079422, | |
| "learning_rate": 2.9651162790697678e-05, | |
| "loss": 0.0033, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 15.30232558139535, | |
| "grad_norm": 2.6470947265625, | |
| "learning_rate": 2.9505813953488377e-05, | |
| "loss": 0.0505, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 15.325581395348838, | |
| "grad_norm": 5.789623737335205, | |
| "learning_rate": 2.9360465116279072e-05, | |
| "loss": 0.0419, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 15.348837209302326, | |
| "grad_norm": 1.2438925504684448, | |
| "learning_rate": 2.921511627906977e-05, | |
| "loss": 0.027, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 15.348837209302326, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8980801776455497, | |
| "eval_loss": 0.36847957968711853, | |
| "eval_precision": 0.8985798152464819, | |
| "eval_recall": 0.8977556646851038, | |
| "eval_runtime": 0.0739, | |
| "eval_samples_per_second": 4125.669, | |
| "eval_steps_per_second": 67.634, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 15.372093023255815, | |
| "grad_norm": 2.8159120082855225, | |
| "learning_rate": 2.9069767441860467e-05, | |
| "loss": 0.0665, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 15.395348837209303, | |
| "grad_norm": 3.529680013656616, | |
| "learning_rate": 2.8924418604651166e-05, | |
| "loss": 0.0238, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 15.418604651162791, | |
| "grad_norm": 0.9973799586296082, | |
| "learning_rate": 2.8779069767441864e-05, | |
| "loss": 0.0051, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 15.44186046511628, | |
| "grad_norm": 2.763955593109131, | |
| "learning_rate": 2.863372093023256e-05, | |
| "loss": 0.0107, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 15.465116279069768, | |
| "grad_norm": 3.6373209953308105, | |
| "learning_rate": 2.848837209302326e-05, | |
| "loss": 0.0466, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 15.488372093023255, | |
| "grad_norm": 3.262023448944092, | |
| "learning_rate": 2.8343023255813954e-05, | |
| "loss": 0.025, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 15.511627906976745, | |
| "grad_norm": 0.7062492370605469, | |
| "learning_rate": 2.8197674418604653e-05, | |
| "loss": 0.0067, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 15.534883720930232, | |
| "grad_norm": 2.9254820346832275, | |
| "learning_rate": 2.805232558139535e-05, | |
| "loss": 0.0394, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 15.55813953488372, | |
| "grad_norm": 1.5290218591690063, | |
| "learning_rate": 2.7906976744186048e-05, | |
| "loss": 0.0268, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 15.581395348837209, | |
| "grad_norm": 3.9581668376922607, | |
| "learning_rate": 2.7761627906976746e-05, | |
| "loss": 0.0683, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 15.604651162790697, | |
| "grad_norm": 3.121964454650879, | |
| "learning_rate": 2.7616279069767442e-05, | |
| "loss": 0.0307, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 15.627906976744185, | |
| "grad_norm": 4.223913669586182, | |
| "learning_rate": 2.747093023255814e-05, | |
| "loss": 0.0778, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 15.651162790697674, | |
| "grad_norm": 1.8996158838272095, | |
| "learning_rate": 2.7325581395348836e-05, | |
| "loss": 0.0149, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 15.674418604651162, | |
| "grad_norm": 2.5367538928985596, | |
| "learning_rate": 2.7180232558139535e-05, | |
| "loss": 0.0238, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 15.69767441860465, | |
| "grad_norm": 1.5632944107055664, | |
| "learning_rate": 2.703488372093023e-05, | |
| "loss": 0.0159, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 15.720930232558139, | |
| "grad_norm": 2.031212329864502, | |
| "learning_rate": 2.688953488372093e-05, | |
| "loss": 0.0265, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 15.744186046511627, | |
| "grad_norm": 1.0653328895568848, | |
| "learning_rate": 2.674418604651163e-05, | |
| "loss": 0.0146, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 15.767441860465116, | |
| "grad_norm": 2.078573226928711, | |
| "learning_rate": 2.6598837209302324e-05, | |
| "loss": 0.0314, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 15.790697674418604, | |
| "grad_norm": 1.004683017730713, | |
| "learning_rate": 2.6453488372093026e-05, | |
| "loss": 0.0079, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 15.813953488372093, | |
| "grad_norm": 0.6423361897468567, | |
| "learning_rate": 2.6308139534883725e-05, | |
| "loss": 0.0265, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 15.813953488372093, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8981460534962133, | |
| "eval_loss": 0.3592439889907837, | |
| "eval_precision": 0.8982758620689655, | |
| "eval_recall": 0.8980356681313001, | |
| "eval_runtime": 0.0743, | |
| "eval_samples_per_second": 4106.888, | |
| "eval_steps_per_second": 67.326, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 15.837209302325581, | |
| "grad_norm": 2.9641597270965576, | |
| "learning_rate": 2.616279069767442e-05, | |
| "loss": 0.0753, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 15.86046511627907, | |
| "grad_norm": 5.338639259338379, | |
| "learning_rate": 2.601744186046512e-05, | |
| "loss": 0.0325, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 15.883720930232558, | |
| "grad_norm": 1.9736541509628296, | |
| "learning_rate": 2.5872093023255818e-05, | |
| "loss": 0.0236, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 15.906976744186046, | |
| "grad_norm": 0.5987362265586853, | |
| "learning_rate": 2.5726744186046514e-05, | |
| "loss": 0.0056, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 15.930232558139535, | |
| "grad_norm": 1.9034494161605835, | |
| "learning_rate": 2.5581395348837212e-05, | |
| "loss": 0.095, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 15.953488372093023, | |
| "grad_norm": 0.7597903609275818, | |
| "learning_rate": 2.5436046511627908e-05, | |
| "loss": 0.0111, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 15.976744186046512, | |
| "grad_norm": 2.3135693073272705, | |
| "learning_rate": 2.5290697674418607e-05, | |
| "loss": 0.0792, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 0.07970131933689117, | |
| "learning_rate": 2.5145348837209302e-05, | |
| "loss": 0.0024, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 16.023255813953487, | |
| "grad_norm": 1.8181146383285522, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.0181, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 16.046511627906977, | |
| "grad_norm": 4.323759078979492, | |
| "learning_rate": 2.48546511627907e-05, | |
| "loss": 0.0386, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 16.069767441860463, | |
| "grad_norm": 3.518137216567993, | |
| "learning_rate": 2.4709302325581396e-05, | |
| "loss": 0.0794, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 16.093023255813954, | |
| "grad_norm": 0.384950190782547, | |
| "learning_rate": 2.4563953488372094e-05, | |
| "loss": 0.0042, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 16.11627906976744, | |
| "grad_norm": 0.4682580232620239, | |
| "learning_rate": 2.441860465116279e-05, | |
| "loss": 0.0078, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 16.13953488372093, | |
| "grad_norm": 1.6920162439346313, | |
| "learning_rate": 2.427325581395349e-05, | |
| "loss": 0.0181, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 16.162790697674417, | |
| "grad_norm": 3.120163679122925, | |
| "learning_rate": 2.4127906976744188e-05, | |
| "loss": 0.0414, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 16.186046511627907, | |
| "grad_norm": 0.8569607734680176, | |
| "learning_rate": 2.3982558139534887e-05, | |
| "loss": 0.0063, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 16.209302325581394, | |
| "grad_norm": 1.5792855024337769, | |
| "learning_rate": 2.3837209302325582e-05, | |
| "loss": 0.0233, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 16.232558139534884, | |
| "grad_norm": 11.798049926757812, | |
| "learning_rate": 2.369186046511628e-05, | |
| "loss": 0.031, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 16.25581395348837, | |
| "grad_norm": 2.0980348587036133, | |
| "learning_rate": 2.354651162790698e-05, | |
| "loss": 0.0345, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 16.27906976744186, | |
| "grad_norm": 1.9684696197509766, | |
| "learning_rate": 2.3401162790697675e-05, | |
| "loss": 0.0109, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 16.27906976744186, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8979217273954117, | |
| "eval_loss": 0.36475130915641785, | |
| "eval_precision": 0.8996056167114501, | |
| "eval_recall": 0.8971956577927114, | |
| "eval_runtime": 0.0743, | |
| "eval_samples_per_second": 4105.148, | |
| "eval_steps_per_second": 67.298, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 16.302325581395348, | |
| "grad_norm": 1.2527034282684326, | |
| "learning_rate": 2.3255813953488374e-05, | |
| "loss": 0.0088, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 16.325581395348838, | |
| "grad_norm": 4.517597198486328, | |
| "learning_rate": 2.311046511627907e-05, | |
| "loss": 0.05, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 16.348837209302324, | |
| "grad_norm": 2.9931998252868652, | |
| "learning_rate": 2.296511627906977e-05, | |
| "loss": 0.0852, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 16.372093023255815, | |
| "grad_norm": 1.5361837148666382, | |
| "learning_rate": 2.2819767441860464e-05, | |
| "loss": 0.0205, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 16.3953488372093, | |
| "grad_norm": 12.71114730834961, | |
| "learning_rate": 2.2674418604651163e-05, | |
| "loss": 0.0643, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 16.41860465116279, | |
| "grad_norm": 0.1732826828956604, | |
| "learning_rate": 2.252906976744186e-05, | |
| "loss": 0.0028, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 16.441860465116278, | |
| "grad_norm": 5.568197250366211, | |
| "learning_rate": 2.238372093023256e-05, | |
| "loss": 0.0147, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 16.46511627906977, | |
| "grad_norm": 3.462963581085205, | |
| "learning_rate": 2.2238372093023256e-05, | |
| "loss": 0.0639, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 16.488372093023255, | |
| "grad_norm": 0.7707027792930603, | |
| "learning_rate": 2.2093023255813955e-05, | |
| "loss": 0.0052, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 16.511627906976745, | |
| "grad_norm": 2.6119136810302734, | |
| "learning_rate": 2.1947674418604654e-05, | |
| "loss": 0.0679, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 16.53488372093023, | |
| "grad_norm": 4.328090190887451, | |
| "learning_rate": 2.180232558139535e-05, | |
| "loss": 0.0336, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 16.558139534883722, | |
| "grad_norm": 3.516861915588379, | |
| "learning_rate": 2.1656976744186048e-05, | |
| "loss": 0.1004, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 16.58139534883721, | |
| "grad_norm": 0.8900988698005676, | |
| "learning_rate": 2.1511627906976744e-05, | |
| "loss": 0.0083, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 16.6046511627907, | |
| "grad_norm": 4.924454689025879, | |
| "learning_rate": 2.1366279069767442e-05, | |
| "loss": 0.1108, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 16.627906976744185, | |
| "grad_norm": 4.6815619468688965, | |
| "learning_rate": 2.1220930232558138e-05, | |
| "loss": 0.0611, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 16.651162790697676, | |
| "grad_norm": 7.625583171844482, | |
| "learning_rate": 2.107558139534884e-05, | |
| "loss": 0.0698, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 16.674418604651162, | |
| "grad_norm": 2.8963229656219482, | |
| "learning_rate": 2.0930232558139536e-05, | |
| "loss": 0.0226, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 16.697674418604652, | |
| "grad_norm": 1.349528193473816, | |
| "learning_rate": 2.0784883720930235e-05, | |
| "loss": 0.0074, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 16.72093023255814, | |
| "grad_norm": 0.9660943746566772, | |
| "learning_rate": 2.0639534883720933e-05, | |
| "loss": 0.0042, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 16.74418604651163, | |
| "grad_norm": 0.205625519156456, | |
| "learning_rate": 2.049418604651163e-05, | |
| "loss": 0.0036, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 16.74418604651163, | |
| "eval_accuracy": 0.9081967213114754, | |
| "eval_f1": 0.9080296372878436, | |
| "eval_loss": 0.3279436528682709, | |
| "eval_precision": 0.9080296372878436, | |
| "eval_recall": 0.9080296372878436, | |
| "eval_runtime": 0.0765, | |
| "eval_samples_per_second": 3988.597, | |
| "eval_steps_per_second": 65.387, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 16.767441860465116, | |
| "grad_norm": 9.84315299987793, | |
| "learning_rate": 2.0348837209302328e-05, | |
| "loss": 0.1307, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 16.790697674418606, | |
| "grad_norm": 11.633748054504395, | |
| "learning_rate": 2.0203488372093023e-05, | |
| "loss": 0.065, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 16.813953488372093, | |
| "grad_norm": 2.6331350803375244, | |
| "learning_rate": 2.0058139534883722e-05, | |
| "loss": 0.0336, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 16.837209302325583, | |
| "grad_norm": 6.747828006744385, | |
| "learning_rate": 1.9912790697674418e-05, | |
| "loss": 0.0478, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 16.86046511627907, | |
| "grad_norm": 1.7129838466644287, | |
| "learning_rate": 1.9767441860465116e-05, | |
| "loss": 0.0403, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 16.88372093023256, | |
| "grad_norm": 1.0863773822784424, | |
| "learning_rate": 1.9622093023255815e-05, | |
| "loss": 0.0105, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 16.906976744186046, | |
| "grad_norm": 1.2210698127746582, | |
| "learning_rate": 1.9476744186046514e-05, | |
| "loss": 0.0114, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 16.930232558139537, | |
| "grad_norm": 3.351280689239502, | |
| "learning_rate": 1.933139534883721e-05, | |
| "loss": 0.0353, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 16.953488372093023, | |
| "grad_norm": 1.9887293577194214, | |
| "learning_rate": 1.918604651162791e-05, | |
| "loss": 0.0208, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 16.97674418604651, | |
| "grad_norm": 0.7091693878173828, | |
| "learning_rate": 1.9040697674418607e-05, | |
| "loss": 0.0074, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "grad_norm": 0.6858041882514954, | |
| "learning_rate": 1.8895348837209303e-05, | |
| "loss": 0.0051, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 17.023255813953487, | |
| "grad_norm": 4.898313045501709, | |
| "learning_rate": 1.8750000000000002e-05, | |
| "loss": 0.1197, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 17.046511627906977, | |
| "grad_norm": 2.349663496017456, | |
| "learning_rate": 1.8604651162790697e-05, | |
| "loss": 0.0244, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 17.069767441860463, | |
| "grad_norm": 2.8203916549682617, | |
| "learning_rate": 1.8459302325581396e-05, | |
| "loss": 0.0801, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 17.093023255813954, | |
| "grad_norm": 4.241973400115967, | |
| "learning_rate": 1.831395348837209e-05, | |
| "loss": 0.0303, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 17.11627906976744, | |
| "grad_norm": 7.765374660491943, | |
| "learning_rate": 1.816860465116279e-05, | |
| "loss": 0.0758, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 17.13953488372093, | |
| "grad_norm": 7.214727401733398, | |
| "learning_rate": 1.802325581395349e-05, | |
| "loss": 0.0461, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 17.162790697674417, | |
| "grad_norm": 7.209427833557129, | |
| "learning_rate": 1.7877906976744188e-05, | |
| "loss": 0.0354, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 17.186046511627907, | |
| "grad_norm": 4.5369744300842285, | |
| "learning_rate": 1.7732558139534887e-05, | |
| "loss": 0.0223, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 17.209302325581394, | |
| "grad_norm": 0.8520297408103943, | |
| "learning_rate": 1.7587209302325583e-05, | |
| "loss": 0.0115, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 17.209302325581394, | |
| "eval_accuracy": 0.8950819672131147, | |
| "eval_f1": 0.8945822172297591, | |
| "eval_loss": 0.35655906796455383, | |
| "eval_precision": 0.8966414996094767, | |
| "eval_recall": 0.8937710002584647, | |
| "eval_runtime": 0.0741, | |
| "eval_samples_per_second": 4118.417, | |
| "eval_steps_per_second": 67.515, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 17.232558139534884, | |
| "grad_norm": 2.845405101776123, | |
| "learning_rate": 1.744186046511628e-05, | |
| "loss": 0.0416, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 17.25581395348837, | |
| "grad_norm": 12.433396339416504, | |
| "learning_rate": 1.7296511627906977e-05, | |
| "loss": 0.0335, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 17.27906976744186, | |
| "grad_norm": 0.4390711486339569, | |
| "learning_rate": 1.7151162790697676e-05, | |
| "loss": 0.0062, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 17.302325581395348, | |
| "grad_norm": 2.333735942840576, | |
| "learning_rate": 1.700581395348837e-05, | |
| "loss": 0.0303, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 17.325581395348838, | |
| "grad_norm": 0.1029660701751709, | |
| "learning_rate": 1.686046511627907e-05, | |
| "loss": 0.0023, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 17.348837209302324, | |
| "grad_norm": 4.543797016143799, | |
| "learning_rate": 1.671511627906977e-05, | |
| "loss": 0.0634, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 17.372093023255815, | |
| "grad_norm": 4.657381534576416, | |
| "learning_rate": 1.6569767441860464e-05, | |
| "loss": 0.0199, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 17.3953488372093, | |
| "grad_norm": 6.3823018074035645, | |
| "learning_rate": 1.6424418604651163e-05, | |
| "loss": 0.1186, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 17.41860465116279, | |
| "grad_norm": 1.975292682647705, | |
| "learning_rate": 1.6279069767441862e-05, | |
| "loss": 0.0948, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 17.441860465116278, | |
| "grad_norm": 4.323884010314941, | |
| "learning_rate": 1.613372093023256e-05, | |
| "loss": 0.1327, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 17.46511627906977, | |
| "grad_norm": 1.2401987314224243, | |
| "learning_rate": 1.5988372093023257e-05, | |
| "loss": 0.0251, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 17.488372093023255, | |
| "grad_norm": 0.8695264458656311, | |
| "learning_rate": 1.5843023255813955e-05, | |
| "loss": 0.005, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 17.511627906976745, | |
| "grad_norm": 2.0256264209747314, | |
| "learning_rate": 1.569767441860465e-05, | |
| "loss": 0.0133, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 17.53488372093023, | |
| "grad_norm": 0.6134868264198303, | |
| "learning_rate": 1.555232558139535e-05, | |
| "loss": 0.0103, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 17.558139534883722, | |
| "grad_norm": 4.545815467834473, | |
| "learning_rate": 1.5406976744186045e-05, | |
| "loss": 0.0191, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 17.58139534883721, | |
| "grad_norm": 3.8483781814575195, | |
| "learning_rate": 1.5261627906976744e-05, | |
| "loss": 0.0311, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 17.6046511627907, | |
| "grad_norm": 2.32814359664917, | |
| "learning_rate": 1.5116279069767441e-05, | |
| "loss": 0.023, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 17.627906976744185, | |
| "grad_norm": 0.3794184625148773, | |
| "learning_rate": 1.4970930232558138e-05, | |
| "loss": 0.0054, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 17.651162790697676, | |
| "grad_norm": 1.3837960958480835, | |
| "learning_rate": 1.4825581395348839e-05, | |
| "loss": 0.0342, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 17.674418604651162, | |
| "grad_norm": 1.8951282501220703, | |
| "learning_rate": 1.4680232558139536e-05, | |
| "loss": 0.0618, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 17.674418604651162, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8980054152598138, | |
| "eval_loss": 0.3718366026878357, | |
| "eval_precision": 0.8990226604393703, | |
| "eval_recall": 0.8974756612389075, | |
| "eval_runtime": 0.0766, | |
| "eval_samples_per_second": 3981.633, | |
| "eval_steps_per_second": 65.273, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 17.697674418604652, | |
| "grad_norm": 1.703765630722046, | |
| "learning_rate": 1.4534883720930233e-05, | |
| "loss": 0.0089, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 17.72093023255814, | |
| "grad_norm": 2.194350004196167, | |
| "learning_rate": 1.4389534883720932e-05, | |
| "loss": 0.0156, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 17.74418604651163, | |
| "grad_norm": 1.5929670333862305, | |
| "learning_rate": 1.424418604651163e-05, | |
| "loss": 0.0093, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 17.767441860465116, | |
| "grad_norm": 1.1527395248413086, | |
| "learning_rate": 1.4098837209302327e-05, | |
| "loss": 0.0087, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 17.790697674418606, | |
| "grad_norm": 0.9619439244270325, | |
| "learning_rate": 1.3953488372093024e-05, | |
| "loss": 0.0101, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 17.813953488372093, | |
| "grad_norm": 2.041372537612915, | |
| "learning_rate": 1.3808139534883721e-05, | |
| "loss": 0.0259, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 17.837209302325583, | |
| "grad_norm": 4.107051849365234, | |
| "learning_rate": 1.3662790697674418e-05, | |
| "loss": 0.0195, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 17.86046511627907, | |
| "grad_norm": 0.8670945763587952, | |
| "learning_rate": 1.3517441860465115e-05, | |
| "loss": 0.0076, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 17.88372093023256, | |
| "grad_norm": 2.5873591899871826, | |
| "learning_rate": 1.3372093023255814e-05, | |
| "loss": 0.0325, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 17.906976744186046, | |
| "grad_norm": 3.630141496658325, | |
| "learning_rate": 1.3226744186046513e-05, | |
| "loss": 0.0261, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 17.930232558139537, | |
| "grad_norm": 0.1693820059299469, | |
| "learning_rate": 1.308139534883721e-05, | |
| "loss": 0.0036, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 17.953488372093023, | |
| "grad_norm": 0.5527888536453247, | |
| "learning_rate": 1.2936046511627909e-05, | |
| "loss": 0.0051, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 17.97674418604651, | |
| "grad_norm": 0.8153883814811707, | |
| "learning_rate": 1.2790697674418606e-05, | |
| "loss": 0.0095, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "grad_norm": 0.2584695518016815, | |
| "learning_rate": 1.2645348837209303e-05, | |
| "loss": 0.0036, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 18.023255813953487, | |
| "grad_norm": 0.5313040018081665, | |
| "learning_rate": 1.25e-05, | |
| "loss": 0.0071, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 18.046511627906977, | |
| "grad_norm": 1.0279793739318848, | |
| "learning_rate": 1.2354651162790698e-05, | |
| "loss": 0.01, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 18.069767441860463, | |
| "grad_norm": 1.0528727769851685, | |
| "learning_rate": 1.2209302325581395e-05, | |
| "loss": 0.0168, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 18.093023255813954, | |
| "grad_norm": 1.2885291576385498, | |
| "learning_rate": 1.2063953488372094e-05, | |
| "loss": 0.0077, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 18.11627906976744, | |
| "grad_norm": 0.3943243622779846, | |
| "learning_rate": 1.1918604651162791e-05, | |
| "loss": 0.0026, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 18.13953488372093, | |
| "grad_norm": 1.1616085767745972, | |
| "learning_rate": 1.177325581395349e-05, | |
| "loss": 0.0061, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 18.13953488372093, | |
| "eval_accuracy": 0.898360655737705, | |
| "eval_f1": 0.8981460534962133, | |
| "eval_loss": 0.37966448068618774, | |
| "eval_precision": 0.8982758620689655, | |
| "eval_recall": 0.8980356681313001, | |
| "eval_runtime": 0.0743, | |
| "eval_samples_per_second": 4106.506, | |
| "eval_steps_per_second": 67.32, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 18.162790697674417, | |
| "grad_norm": 0.11722904443740845, | |
| "learning_rate": 1.1627906976744187e-05, | |
| "loss": 0.0023, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 18.186046511627907, | |
| "grad_norm": 1.2253535985946655, | |
| "learning_rate": 1.1482558139534884e-05, | |
| "loss": 0.0182, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 18.209302325581394, | |
| "grad_norm": 0.4131571054458618, | |
| "learning_rate": 1.1337209302325581e-05, | |
| "loss": 0.0038, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 18.232558139534884, | |
| "grad_norm": 0.5466241240501404, | |
| "learning_rate": 1.119186046511628e-05, | |
| "loss": 0.0033, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 18.25581395348837, | |
| "grad_norm": 0.8692965507507324, | |
| "learning_rate": 1.1046511627906977e-05, | |
| "loss": 0.0079, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 18.27906976744186, | |
| "grad_norm": 3.9168105125427246, | |
| "learning_rate": 1.0901162790697675e-05, | |
| "loss": 0.0861, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 18.302325581395348, | |
| "grad_norm": 1.2862012386322021, | |
| "learning_rate": 1.0755813953488372e-05, | |
| "loss": 0.0081, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 18.325581395348838, | |
| "grad_norm": 6.316368103027344, | |
| "learning_rate": 1.0610465116279069e-05, | |
| "loss": 0.044, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 18.348837209302324, | |
| "grad_norm": 7.892980098724365, | |
| "learning_rate": 1.0465116279069768e-05, | |
| "loss": 0.1037, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 18.372093023255815, | |
| "grad_norm": 7.4783172607421875, | |
| "learning_rate": 1.0319767441860467e-05, | |
| "loss": 0.0339, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 18.3953488372093, | |
| "grad_norm": 0.7634483575820923, | |
| "learning_rate": 1.0174418604651164e-05, | |
| "loss": 0.0039, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 18.41860465116279, | |
| "grad_norm": 3.145096778869629, | |
| "learning_rate": 1.0029069767441861e-05, | |
| "loss": 0.0739, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 18.441860465116278, | |
| "grad_norm": 6.179558277130127, | |
| "learning_rate": 9.883720930232558e-06, | |
| "loss": 0.0333, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 18.46511627906977, | |
| "grad_norm": 0.10819542407989502, | |
| "learning_rate": 9.738372093023257e-06, | |
| "loss": 0.0023, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 18.488372093023255, | |
| "grad_norm": 11.038155555725098, | |
| "learning_rate": 9.593023255813954e-06, | |
| "loss": 0.0816, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 18.511627906976745, | |
| "grad_norm": 4.89321756362915, | |
| "learning_rate": 9.447674418604651e-06, | |
| "loss": 0.0669, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 18.53488372093023, | |
| "grad_norm": 0.6758410930633545, | |
| "learning_rate": 9.302325581395349e-06, | |
| "loss": 0.0124, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 18.558139534883722, | |
| "grad_norm": 2.737079381942749, | |
| "learning_rate": 9.156976744186046e-06, | |
| "loss": 0.0134, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 18.58139534883721, | |
| "grad_norm": 5.82487154006958, | |
| "learning_rate": 9.011627906976745e-06, | |
| "loss": 0.0826, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 18.6046511627907, | |
| "grad_norm": 0.12569262087345123, | |
| "learning_rate": 8.866279069767444e-06, | |
| "loss": 0.0021, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 18.6046511627907, | |
| "eval_accuracy": 0.8950819672131147, | |
| "eval_f1": 0.8945822172297591, | |
| "eval_loss": 0.41386428475379944, | |
| "eval_precision": 0.8966414996094767, | |
| "eval_recall": 0.8937710002584647, | |
| "eval_runtime": 0.074, | |
| "eval_samples_per_second": 4121.655, | |
| "eval_steps_per_second": 67.568, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 18.627906976744185, | |
| "grad_norm": 7.690613269805908, | |
| "learning_rate": 8.72093023255814e-06, | |
| "loss": 0.1334, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 18.651162790697676, | |
| "grad_norm": 7.942806720733643, | |
| "learning_rate": 8.575581395348838e-06, | |
| "loss": 0.0349, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 18.674418604651162, | |
| "grad_norm": 1.4096626043319702, | |
| "learning_rate": 8.430232558139535e-06, | |
| "loss": 0.0086, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 18.697674418604652, | |
| "grad_norm": 2.184845209121704, | |
| "learning_rate": 8.284883720930232e-06, | |
| "loss": 0.021, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 18.72093023255814, | |
| "grad_norm": 5.123547554016113, | |
| "learning_rate": 8.139534883720931e-06, | |
| "loss": 0.0714, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 18.74418604651163, | |
| "grad_norm": 2.141044855117798, | |
| "learning_rate": 7.994186046511628e-06, | |
| "loss": 0.0227, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 18.767441860465116, | |
| "grad_norm": 0.7044129967689514, | |
| "learning_rate": 7.848837209302325e-06, | |
| "loss": 0.0064, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 18.790697674418606, | |
| "grad_norm": 0.9102310538291931, | |
| "learning_rate": 7.703488372093023e-06, | |
| "loss": 0.0056, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 18.813953488372093, | |
| "grad_norm": 6.990550518035889, | |
| "learning_rate": 7.558139534883721e-06, | |
| "loss": 0.0359, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 18.837209302325583, | |
| "grad_norm": 0.36822423338890076, | |
| "learning_rate": 7.4127906976744195e-06, | |
| "loss": 0.0036, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 18.86046511627907, | |
| "grad_norm": 1.0232497453689575, | |
| "learning_rate": 7.267441860465117e-06, | |
| "loss": 0.0077, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 18.88372093023256, | |
| "grad_norm": 2.3726754188537598, | |
| "learning_rate": 7.122093023255815e-06, | |
| "loss": 0.0625, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 18.906976744186046, | |
| "grad_norm": 1.6161715984344482, | |
| "learning_rate": 6.976744186046512e-06, | |
| "loss": 0.015, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 18.930232558139537, | |
| "grad_norm": 1.230698823928833, | |
| "learning_rate": 6.831395348837209e-06, | |
| "loss": 0.0595, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 18.953488372093023, | |
| "grad_norm": 4.402613162994385, | |
| "learning_rate": 6.686046511627907e-06, | |
| "loss": 0.0847, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 18.97674418604651, | |
| "grad_norm": 6.6045966148376465, | |
| "learning_rate": 6.540697674418605e-06, | |
| "loss": 0.036, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "grad_norm": 0.6431266069412231, | |
| "learning_rate": 6.395348837209303e-06, | |
| "loss": 0.0043, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 19.023255813953487, | |
| "grad_norm": 0.4433208703994751, | |
| "learning_rate": 6.25e-06, | |
| "loss": 0.0032, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 19.046511627906977, | |
| "grad_norm": 3.6785125732421875, | |
| "learning_rate": 6.1046511627906975e-06, | |
| "loss": 0.1244, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 19.069767441860463, | |
| "grad_norm": 2.5291504859924316, | |
| "learning_rate": 5.9593023255813955e-06, | |
| "loss": 0.0071, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 19.069767441860463, | |
| "eval_accuracy": 0.9049180327868852, | |
| "eval_f1": 0.9046556500555143, | |
| "eval_loss": 0.3841802775859833, | |
| "eval_precision": 0.9051627384960719, | |
| "eval_recall": 0.9043249763074007, | |
| "eval_runtime": 0.0744, | |
| "eval_samples_per_second": 4097.049, | |
| "eval_steps_per_second": 67.165, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 19.093023255813954, | |
| "grad_norm": 8.134071350097656, | |
| "learning_rate": 5.8139534883720935e-06, | |
| "loss": 0.0308, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 19.11627906976744, | |
| "grad_norm": 0.23603279888629913, | |
| "learning_rate": 5.668604651162791e-06, | |
| "loss": 0.0028, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 19.13953488372093, | |
| "grad_norm": 5.337473392486572, | |
| "learning_rate": 5.523255813953489e-06, | |
| "loss": 0.034, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 19.162790697674417, | |
| "grad_norm": 4.512674808502197, | |
| "learning_rate": 5.377906976744186e-06, | |
| "loss": 0.0775, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 19.186046511627907, | |
| "grad_norm": 1.7216310501098633, | |
| "learning_rate": 5.232558139534884e-06, | |
| "loss": 0.0056, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 19.209302325581394, | |
| "grad_norm": 1.9352085590362549, | |
| "learning_rate": 5.087209302325582e-06, | |
| "loss": 0.0423, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 19.232558139534884, | |
| "grad_norm": 10.692374229431152, | |
| "learning_rate": 4.941860465116279e-06, | |
| "loss": 0.0782, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 19.25581395348837, | |
| "grad_norm": 0.10650653392076492, | |
| "learning_rate": 4.796511627906977e-06, | |
| "loss": 0.0017, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 19.27906976744186, | |
| "grad_norm": 0.14889536798000336, | |
| "learning_rate": 4.651162790697674e-06, | |
| "loss": 0.0024, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 19.302325581395348, | |
| "grad_norm": 5.617801666259766, | |
| "learning_rate": 4.505813953488372e-06, | |
| "loss": 0.0121, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 19.325581395348838, | |
| "grad_norm": 2.859057903289795, | |
| "learning_rate": 4.36046511627907e-06, | |
| "loss": 0.034, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 19.348837209302324, | |
| "grad_norm": 0.27750277519226074, | |
| "learning_rate": 4.2151162790697675e-06, | |
| "loss": 0.0044, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 19.372093023255815, | |
| "grad_norm": 0.04883375018835068, | |
| "learning_rate": 4.0697674418604655e-06, | |
| "loss": 0.0018, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 19.3953488372093, | |
| "grad_norm": 1.9595685005187988, | |
| "learning_rate": 3.924418604651163e-06, | |
| "loss": 0.011, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 19.41860465116279, | |
| "grad_norm": 1.273353099822998, | |
| "learning_rate": 3.7790697674418603e-06, | |
| "loss": 0.0214, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 19.441860465116278, | |
| "grad_norm": 6.320394039154053, | |
| "learning_rate": 3.6337209302325583e-06, | |
| "loss": 0.081, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 19.46511627906977, | |
| "grad_norm": 2.93029522895813, | |
| "learning_rate": 3.488372093023256e-06, | |
| "loss": 0.0256, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 19.488372093023255, | |
| "grad_norm": 0.4745703339576721, | |
| "learning_rate": 3.3430232558139535e-06, | |
| "loss": 0.0033, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 19.511627906976745, | |
| "grad_norm": 4.16102933883667, | |
| "learning_rate": 3.1976744186046516e-06, | |
| "loss": 0.0482, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 19.53488372093023, | |
| "grad_norm": 0.2739373743534088, | |
| "learning_rate": 3.0523255813953487e-06, | |
| "loss": 0.0045, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 19.53488372093023, | |
| "eval_accuracy": 0.9049180327868852, | |
| "eval_f1": 0.9046556500555143, | |
| "eval_loss": 0.38903769850730896, | |
| "eval_precision": 0.9051627384960719, | |
| "eval_recall": 0.9043249763074007, | |
| "eval_runtime": 0.0754, | |
| "eval_samples_per_second": 4042.913, | |
| "eval_steps_per_second": 66.277, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 19.558139534883722, | |
| "grad_norm": 0.87278813123703, | |
| "learning_rate": 2.9069767441860468e-06, | |
| "loss": 0.0116, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 19.58139534883721, | |
| "grad_norm": 0.7536394596099854, | |
| "learning_rate": 2.7616279069767444e-06, | |
| "loss": 0.0036, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 19.6046511627907, | |
| "grad_norm": 4.64411735534668, | |
| "learning_rate": 2.616279069767442e-06, | |
| "loss": 0.0712, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 19.627906976744185, | |
| "grad_norm": 0.5402886271476746, | |
| "learning_rate": 2.4709302325581396e-06, | |
| "loss": 0.0079, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 19.651162790697676, | |
| "grad_norm": 0.591332733631134, | |
| "learning_rate": 2.325581395348837e-06, | |
| "loss": 0.0028, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 19.674418604651162, | |
| "grad_norm": 1.511197805404663, | |
| "learning_rate": 2.180232558139535e-06, | |
| "loss": 0.0216, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 19.697674418604652, | |
| "grad_norm": 2.847646951675415, | |
| "learning_rate": 2.0348837209302328e-06, | |
| "loss": 0.0246, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 19.72093023255814, | |
| "grad_norm": 0.33959656953811646, | |
| "learning_rate": 1.8895348837209302e-06, | |
| "loss": 0.0034, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 19.74418604651163, | |
| "grad_norm": 1.8157447576522827, | |
| "learning_rate": 1.744186046511628e-06, | |
| "loss": 0.0436, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 19.767441860465116, | |
| "grad_norm": 0.3644404113292694, | |
| "learning_rate": 1.5988372093023258e-06, | |
| "loss": 0.0038, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 19.790697674418606, | |
| "grad_norm": 2.5243899822235107, | |
| "learning_rate": 1.4534883720930234e-06, | |
| "loss": 0.0115, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 19.813953488372093, | |
| "grad_norm": 5.748671054840088, | |
| "learning_rate": 1.308139534883721e-06, | |
| "loss": 0.04, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 19.837209302325583, | |
| "grad_norm": 0.5189786553382874, | |
| "learning_rate": 1.1627906976744186e-06, | |
| "loss": 0.0049, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 19.86046511627907, | |
| "grad_norm": 2.05354380607605, | |
| "learning_rate": 1.0174418604651164e-06, | |
| "loss": 0.0591, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 19.88372093023256, | |
| "grad_norm": 2.982142925262451, | |
| "learning_rate": 8.72093023255814e-07, | |
| "loss": 0.0224, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 19.906976744186046, | |
| "grad_norm": 2.0666937828063965, | |
| "learning_rate": 7.267441860465117e-07, | |
| "loss": 0.0217, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 19.930232558139537, | |
| "grad_norm": 4.0683722496032715, | |
| "learning_rate": 5.813953488372093e-07, | |
| "loss": 0.108, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 19.953488372093023, | |
| "grad_norm": 10.890898704528809, | |
| "learning_rate": 4.36046511627907e-07, | |
| "loss": 0.069, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 19.97674418604651, | |
| "grad_norm": 0.7343159914016724, | |
| "learning_rate": 2.9069767441860464e-07, | |
| "loss": 0.0071, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 1.6608028411865234, | |
| "learning_rate": 1.4534883720930232e-07, | |
| "loss": 0.0411, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 0.8950819672131147, | |
| "eval_f1": 0.8946732280065612, | |
| "eval_loss": 0.3955073356628418, | |
| "eval_precision": 0.895995670995671, | |
| "eval_recall": 0.894051003704661, | |
| "eval_runtime": 0.076, | |
| "eval_samples_per_second": 4014.381, | |
| "eval_steps_per_second": 65.81, | |
| "step": 860 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 860, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 20, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 68001814271952.0, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |