{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 20, "global_step": 1078, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "eval_accuracy": 0.738359201773836, "eval_f1": 0.23376623376623376, "eval_loss": 0.5994226336479187, "eval_precision": 0.6545454545454545, "eval_recall": 0.1422924901185771, "eval_runtime": 47.8594, "eval_samples_per_second": 5.767, "eval_steps_per_second": 0.188, "step": 0 }, { "epoch": 0.0009276437847866419, "grad_norm": 2.113847255706787, "learning_rate": 1.8518518518518518e-07, "loss": 0.6464, "step": 1 }, { "epoch": 0.0018552875695732839, "grad_norm": 2.337275505065918, "learning_rate": 3.7037037037037036e-07, "loss": 0.7325, "step": 2 }, { "epoch": 0.0027829313543599257, "grad_norm": 2.0731401443481445, "learning_rate": 5.555555555555555e-07, "loss": 0.6441, "step": 3 }, { "epoch": 0.0037105751391465678, "grad_norm": 2.075726270675659, "learning_rate": 7.407407407407407e-07, "loss": 0.6722, "step": 4 }, { "epoch": 0.00463821892393321, "grad_norm": 2.2193410396575928, "learning_rate": 9.259259259259259e-07, "loss": 0.6676, "step": 5 }, { "epoch": 0.0055658627087198514, "grad_norm": 2.2222299575805664, "learning_rate": 1.111111111111111e-06, "loss": 0.5932, "step": 6 }, { "epoch": 0.006493506493506494, "grad_norm": 2.5532214641571045, "learning_rate": 1.2962962962962962e-06, "loss": 0.6558, "step": 7 }, { "epoch": 0.0074211502782931356, "grad_norm": 1.7259494066238403, "learning_rate": 1.4814814814814815e-06, "loss": 0.5976, "step": 8 }, { "epoch": 0.008348794063079777, "grad_norm": 2.020770788192749, "learning_rate": 1.6666666666666667e-06, "loss": 0.6576, "step": 9 }, { "epoch": 0.00927643784786642, "grad_norm": 2.5885586738586426, "learning_rate": 1.8518518518518519e-06, "loss": 0.6656, "step": 10 }, { "epoch": 0.01020408163265306, "grad_norm": 1.901673674583435, "learning_rate": 2.037037037037037e-06, "loss": 0.6293, "step": 11 }, { "epoch": 0.011131725417439703, "grad_norm": 1.867441177368164, "learning_rate": 2.222222222222222e-06, "loss": 0.5698, "step": 12 }, { "epoch": 0.012059369202226345, "grad_norm": 2.405125141143799, "learning_rate": 2.4074074074074075e-06, "loss": 0.5984, "step": 13 }, { "epoch": 0.012987012987012988, "grad_norm": 2.0719292163848877, "learning_rate": 2.5925925925925925e-06, "loss": 0.5933, "step": 14 }, { "epoch": 0.013914656771799629, "grad_norm": 2.2819459438323975, "learning_rate": 2.7777777777777783e-06, "loss": 0.6439, "step": 15 }, { "epoch": 0.014842300556586271, "grad_norm": 2.3114988803863525, "learning_rate": 2.962962962962963e-06, "loss": 0.6118, "step": 16 }, { "epoch": 0.015769944341372914, "grad_norm": 2.1615548133850098, "learning_rate": 3.1481481481481483e-06, "loss": 0.5185, "step": 17 }, { "epoch": 0.016697588126159554, "grad_norm": 2.3003029823303223, "learning_rate": 3.3333333333333333e-06, "loss": 0.6267, "step": 18 }, { "epoch": 0.017625231910946195, "grad_norm": 2.284081220626831, "learning_rate": 3.5185185185185187e-06, "loss": 0.6659, "step": 19 }, { "epoch": 0.01855287569573284, "grad_norm": 1.8381303548812866, "learning_rate": 3.7037037037037037e-06, "loss": 0.572, "step": 20 }, { "epoch": 0.01855287569573284, "eval_accuracy": 0.7394678492239468, "eval_f1": 0.22950819672131148, "eval_loss": 0.5945659279823303, "eval_precision": 0.6730769230769231, "eval_recall": 0.1383399209486166, "eval_runtime": 47.6156, "eval_samples_per_second": 5.796, "eval_steps_per_second": 0.189, "step": 20 }, { "epoch": 0.01948051948051948, "grad_norm": 2.036238670349121, "learning_rate": 3.88888888888889e-06, "loss": 0.6409, "step": 21 }, { "epoch": 0.02040816326530612, "grad_norm": 1.9110214710235596, "learning_rate": 4.074074074074074e-06, "loss": 0.6008, "step": 22 }, { "epoch": 0.021335807050092765, "grad_norm": 2.4081571102142334, "learning_rate": 4.2592592592592596e-06, "loss": 0.632, "step": 23 }, { "epoch": 0.022263450834879406, "grad_norm": 1.9782453775405884, "learning_rate": 4.444444444444444e-06, "loss": 0.6474, "step": 24 }, { "epoch": 0.023191094619666047, "grad_norm": 1.7934902906417847, "learning_rate": 4.62962962962963e-06, "loss": 0.5803, "step": 25 }, { "epoch": 0.02411873840445269, "grad_norm": 2.6915359497070312, "learning_rate": 4.814814814814815e-06, "loss": 0.6505, "step": 26 }, { "epoch": 0.02504638218923933, "grad_norm": 1.806140422821045, "learning_rate": 5e-06, "loss": 0.5283, "step": 27 }, { "epoch": 0.025974025974025976, "grad_norm": 1.7927738428115845, "learning_rate": 5.185185185185185e-06, "loss": 0.5696, "step": 28 }, { "epoch": 0.026901669758812616, "grad_norm": 1.7386640310287476, "learning_rate": 5.370370370370371e-06, "loss": 0.4902, "step": 29 }, { "epoch": 0.027829313543599257, "grad_norm": 2.015052318572998, "learning_rate": 5.555555555555557e-06, "loss": 0.6123, "step": 30 }, { "epoch": 0.0287569573283859, "grad_norm": 1.620802402496338, "learning_rate": 5.740740740740741e-06, "loss": 0.5611, "step": 31 }, { "epoch": 0.029684601113172542, "grad_norm": 1.8895862102508545, "learning_rate": 5.925925925925926e-06, "loss": 0.6024, "step": 32 }, { "epoch": 0.030612244897959183, "grad_norm": 2.030715227127075, "learning_rate": 6.111111111111112e-06, "loss": 0.6245, "step": 33 }, { "epoch": 0.03153988868274583, "grad_norm": 2.208017349243164, "learning_rate": 6.296296296296297e-06, "loss": 0.5331, "step": 34 }, { "epoch": 0.032467532467532464, "grad_norm": 2.373842239379883, "learning_rate": 6.481481481481482e-06, "loss": 0.638, "step": 35 }, { "epoch": 0.03339517625231911, "grad_norm": 2.170443296432495, "learning_rate": 6.666666666666667e-06, "loss": 0.563, "step": 36 }, { "epoch": 0.03432282003710575, "grad_norm": 1.4894213676452637, "learning_rate": 6.851851851851853e-06, "loss": 0.5338, "step": 37 }, { "epoch": 0.03525046382189239, "grad_norm": 1.9410276412963867, "learning_rate": 7.0370370370370375e-06, "loss": 0.5609, "step": 38 }, { "epoch": 0.036178107606679034, "grad_norm": 2.315661668777466, "learning_rate": 7.222222222222223e-06, "loss": 0.4866, "step": 39 }, { "epoch": 0.03710575139146568, "grad_norm": 1.8483999967575073, "learning_rate": 7.4074074074074075e-06, "loss": 0.5829, "step": 40 }, { "epoch": 0.03710575139146568, "eval_accuracy": 0.7494456762749445, "eval_f1": 0.2848101265822785, "eval_loss": 0.5677995681762695, "eval_precision": 0.7142857142857143, "eval_recall": 0.17786561264822134, "eval_runtime": 46.6917, "eval_samples_per_second": 5.911, "eval_steps_per_second": 0.193, "step": 40 }, { "epoch": 0.038033395176252316, "grad_norm": 2.0759365558624268, "learning_rate": 7.592592592592594e-06, "loss": 0.5541, "step": 41 }, { "epoch": 0.03896103896103896, "grad_norm": 1.7480107545852661, "learning_rate": 7.77777777777778e-06, "loss": 0.5236, "step": 42 }, { "epoch": 0.039888682745825604, "grad_norm": 1.8141703605651855, "learning_rate": 7.962962962962963e-06, "loss": 0.519, "step": 43 }, { "epoch": 0.04081632653061224, "grad_norm": 1.8323687314987183, "learning_rate": 8.148148148148148e-06, "loss": 0.5472, "step": 44 }, { "epoch": 0.041743970315398886, "grad_norm": 1.6657449007034302, "learning_rate": 8.333333333333334e-06, "loss": 0.4735, "step": 45 }, { "epoch": 0.04267161410018553, "grad_norm": 2.900596857070923, "learning_rate": 8.518518518518519e-06, "loss": 0.5968, "step": 46 }, { "epoch": 0.04359925788497217, "grad_norm": 1.9013563394546509, "learning_rate": 8.703703703703705e-06, "loss": 0.5436, "step": 47 }, { "epoch": 0.04452690166975881, "grad_norm": 1.7270978689193726, "learning_rate": 8.888888888888888e-06, "loss": 0.4987, "step": 48 }, { "epoch": 0.045454545454545456, "grad_norm": 1.4679995775222778, "learning_rate": 9.074074074074075e-06, "loss": 0.4405, "step": 49 }, { "epoch": 0.04638218923933209, "grad_norm": 2.332545042037964, "learning_rate": 9.25925925925926e-06, "loss": 0.5181, "step": 50 }, { "epoch": 0.04730983302411874, "grad_norm": 2.0095181465148926, "learning_rate": 9.444444444444445e-06, "loss": 0.5561, "step": 51 }, { "epoch": 0.04823747680890538, "grad_norm": 2.628974199295044, "learning_rate": 9.62962962962963e-06, "loss": 0.5791, "step": 52 }, { "epoch": 0.04916512059369202, "grad_norm": 1.918746829032898, "learning_rate": 9.814814814814815e-06, "loss": 0.5171, "step": 53 }, { "epoch": 0.05009276437847866, "grad_norm": 1.907528281211853, "learning_rate": 1e-05, "loss": 0.5104, "step": 54 }, { "epoch": 0.05102040816326531, "grad_norm": 1.70274817943573, "learning_rate": 1.0185185185185186e-05, "loss": 0.5104, "step": 55 }, { "epoch": 0.05194805194805195, "grad_norm": 2.177898406982422, "learning_rate": 1.037037037037037e-05, "loss": 0.4683, "step": 56 }, { "epoch": 0.05287569573283859, "grad_norm": 2.1687746047973633, "learning_rate": 1.0555555555555557e-05, "loss": 0.4888, "step": 57 }, { "epoch": 0.05380333951762523, "grad_norm": 3.19343900680542, "learning_rate": 1.0740740740740742e-05, "loss": 0.5476, "step": 58 }, { "epoch": 0.05473098330241188, "grad_norm": 2.7876338958740234, "learning_rate": 1.0925925925925926e-05, "loss": 0.4876, "step": 59 }, { "epoch": 0.055658627087198514, "grad_norm": 2.028104066848755, "learning_rate": 1.1111111111111113e-05, "loss": 0.4808, "step": 60 }, { "epoch": 0.055658627087198514, "eval_accuracy": 0.7694013303769401, "eval_f1": 0.45549738219895286, "eval_loss": 0.512853741645813, "eval_precision": 0.6744186046511628, "eval_recall": 0.3438735177865613, "eval_runtime": 46.8717, "eval_samples_per_second": 5.888, "eval_steps_per_second": 0.192, "step": 60 }, { "epoch": 0.05658627087198516, "grad_norm": 2.1539430618286133, "learning_rate": 1.1296296296296297e-05, "loss": 0.5098, "step": 61 }, { "epoch": 0.0575139146567718, "grad_norm": 2.1614112854003906, "learning_rate": 1.1481481481481482e-05, "loss": 0.4424, "step": 62 }, { "epoch": 0.05844155844155844, "grad_norm": 2.3000025749206543, "learning_rate": 1.1666666666666668e-05, "loss": 0.4228, "step": 63 }, { "epoch": 0.059369202226345084, "grad_norm": 2.695854663848877, "learning_rate": 1.1851851851851852e-05, "loss": 0.4409, "step": 64 }, { "epoch": 0.06029684601113173, "grad_norm": 2.0516905784606934, "learning_rate": 1.2037037037037039e-05, "loss": 0.5061, "step": 65 }, { "epoch": 0.061224489795918366, "grad_norm": 2.1683146953582764, "learning_rate": 1.2222222222222224e-05, "loss": 0.4346, "step": 66 }, { "epoch": 0.06215213358070501, "grad_norm": 1.7861865758895874, "learning_rate": 1.2407407407407408e-05, "loss": 0.4451, "step": 67 }, { "epoch": 0.06307977736549165, "grad_norm": 2.505061626434326, "learning_rate": 1.2592592592592593e-05, "loss": 0.4545, "step": 68 }, { "epoch": 0.0640074211502783, "grad_norm": 2.663344621658325, "learning_rate": 1.2777777777777777e-05, "loss": 0.4741, "step": 69 }, { "epoch": 0.06493506493506493, "grad_norm": 2.4229860305786133, "learning_rate": 1.2962962962962964e-05, "loss": 0.4564, "step": 70 }, { "epoch": 0.06586270871985157, "grad_norm": 1.9748098850250244, "learning_rate": 1.314814814814815e-05, "loss": 0.4832, "step": 71 }, { "epoch": 0.06679035250463822, "grad_norm": 2.0413177013397217, "learning_rate": 1.3333333333333333e-05, "loss": 0.4377, "step": 72 }, { "epoch": 0.06771799628942486, "grad_norm": 2.965312957763672, "learning_rate": 1.351851851851852e-05, "loss": 0.3847, "step": 73 }, { "epoch": 0.0686456400742115, "grad_norm": 3.3258345127105713, "learning_rate": 1.3703703703703706e-05, "loss": 0.441, "step": 74 }, { "epoch": 0.06957328385899815, "grad_norm": 4.911780834197998, "learning_rate": 1.388888888888889e-05, "loss": 0.5277, "step": 75 }, { "epoch": 0.07050092764378478, "grad_norm": 3.556264638900757, "learning_rate": 1.4074074074074075e-05, "loss": 0.5102, "step": 76 }, { "epoch": 0.07142857142857142, "grad_norm": 2.313514232635498, "learning_rate": 1.4259259259259259e-05, "loss": 0.3874, "step": 77 }, { "epoch": 0.07235621521335807, "grad_norm": 2.3410682678222656, "learning_rate": 1.4444444444444446e-05, "loss": 0.3794, "step": 78 }, { "epoch": 0.07328385899814471, "grad_norm": 1.974550724029541, "learning_rate": 1.4629629629629631e-05, "loss": 0.3691, "step": 79 }, { "epoch": 0.07421150278293136, "grad_norm": 3.7399516105651855, "learning_rate": 1.4814814814814815e-05, "loss": 0.498, "step": 80 }, { "epoch": 0.07421150278293136, "eval_accuracy": 0.7804878048780488, "eval_f1": 0.5217391304347826, "eval_loss": 0.46577775478363037, "eval_precision": 0.6708074534161491, "eval_recall": 0.4268774703557312, "eval_runtime": 47.3211, "eval_samples_per_second": 5.832, "eval_steps_per_second": 0.19, "step": 80 }, { "epoch": 0.075139146567718, "grad_norm": 4.030659198760986, "learning_rate": 1.5000000000000002e-05, "loss": 0.4012, "step": 81 }, { "epoch": 0.07606679035250463, "grad_norm": 4.128874778747559, "learning_rate": 1.5185185185185187e-05, "loss": 0.4411, "step": 82 }, { "epoch": 0.07699443413729128, "grad_norm": 5.405181884765625, "learning_rate": 1.537037037037037e-05, "loss": 0.3886, "step": 83 }, { "epoch": 0.07792207792207792, "grad_norm": 2.473198175430298, "learning_rate": 1.555555555555556e-05, "loss": 0.3348, "step": 84 }, { "epoch": 0.07884972170686456, "grad_norm": 3.9382104873657227, "learning_rate": 1.5740740740740744e-05, "loss": 0.3732, "step": 85 }, { "epoch": 0.07977736549165121, "grad_norm": 2.209444761276245, "learning_rate": 1.5925925925925926e-05, "loss": 0.3247, "step": 86 }, { "epoch": 0.08070500927643785, "grad_norm": 4.216385364532471, "learning_rate": 1.6111111111111115e-05, "loss": 0.4295, "step": 87 }, { "epoch": 0.08163265306122448, "grad_norm": 4.730212211608887, "learning_rate": 1.6296296296296297e-05, "loss": 0.3214, "step": 88 }, { "epoch": 0.08256029684601113, "grad_norm": 2.821192741394043, "learning_rate": 1.6481481481481482e-05, "loss": 0.3864, "step": 89 }, { "epoch": 0.08348794063079777, "grad_norm": 2.9178500175476074, "learning_rate": 1.6666666666666667e-05, "loss": 0.3257, "step": 90 }, { "epoch": 0.08441558441558442, "grad_norm": 4.074183464050293, "learning_rate": 1.6851851851851853e-05, "loss": 0.3916, "step": 91 }, { "epoch": 0.08534322820037106, "grad_norm": 2.721669912338257, "learning_rate": 1.7037037037037038e-05, "loss": 0.3153, "step": 92 }, { "epoch": 0.0862708719851577, "grad_norm": 4.657402992248535, "learning_rate": 1.7222222222222224e-05, "loss": 0.4462, "step": 93 }, { "epoch": 0.08719851576994433, "grad_norm": 2.9709548950195312, "learning_rate": 1.740740740740741e-05, "loss": 0.3124, "step": 94 }, { "epoch": 0.08812615955473098, "grad_norm": 5.8010711669921875, "learning_rate": 1.7592592592592595e-05, "loss": 0.2713, "step": 95 }, { "epoch": 0.08905380333951762, "grad_norm": 3.4263553619384766, "learning_rate": 1.7777777777777777e-05, "loss": 0.3161, "step": 96 }, { "epoch": 0.08998144712430427, "grad_norm": 3.680510997772217, "learning_rate": 1.7962962962962965e-05, "loss": 0.3464, "step": 97 }, { "epoch": 0.09090909090909091, "grad_norm": 5.526573657989502, "learning_rate": 1.814814814814815e-05, "loss": 0.2899, "step": 98 }, { "epoch": 0.09183673469387756, "grad_norm": 3.2282707691192627, "learning_rate": 1.8333333333333333e-05, "loss": 0.2088, "step": 99 }, { "epoch": 0.09276437847866419, "grad_norm": 3.8928277492523193, "learning_rate": 1.851851851851852e-05, "loss": 0.2531, "step": 100 }, { "epoch": 0.09276437847866419, "eval_accuracy": 0.8015521064301552, "eval_f1": 0.5665859564164649, "eval_loss": 0.4834875166416168, "eval_precision": 0.73125, "eval_recall": 0.4624505928853755, "eval_runtime": 48.1977, "eval_samples_per_second": 5.726, "eval_steps_per_second": 0.187, "step": 100 }, { "epoch": 0.09369202226345083, "grad_norm": 4.2124924659729, "learning_rate": 1.8703703703703707e-05, "loss": 0.2609, "step": 101 }, { "epoch": 0.09461966604823747, "grad_norm": 5.753302097320557, "learning_rate": 1.888888888888889e-05, "loss": 0.3886, "step": 102 }, { "epoch": 0.09554730983302412, "grad_norm": 4.0034050941467285, "learning_rate": 1.9074074074074075e-05, "loss": 0.2714, "step": 103 }, { "epoch": 0.09647495361781076, "grad_norm": 7.041585922241211, "learning_rate": 1.925925925925926e-05, "loss": 0.3665, "step": 104 }, { "epoch": 0.09740259740259741, "grad_norm": 5.10760498046875, "learning_rate": 1.9444444444444445e-05, "loss": 0.3047, "step": 105 }, { "epoch": 0.09833024118738404, "grad_norm": 7.445008754730225, "learning_rate": 1.962962962962963e-05, "loss": 0.2991, "step": 106 }, { "epoch": 0.09925788497217068, "grad_norm": 5.59462833404541, "learning_rate": 1.9814814814814816e-05, "loss": 0.3124, "step": 107 }, { "epoch": 0.10018552875695733, "grad_norm": 4.407949447631836, "learning_rate": 2e-05, "loss": 0.2371, "step": 108 }, { "epoch": 0.10111317254174397, "grad_norm": 7.777821063995361, "learning_rate": 1.999994755236596e-05, "loss": 0.4208, "step": 109 }, { "epoch": 0.10204081632653061, "grad_norm": 9.282930374145508, "learning_rate": 1.999979021001399e-05, "loss": 0.2394, "step": 110 }, { "epoch": 0.10296846011131726, "grad_norm": 3.985445976257324, "learning_rate": 1.999952797459453e-05, "loss": 0.3059, "step": 111 }, { "epoch": 0.1038961038961039, "grad_norm": 8.832866668701172, "learning_rate": 1.999916084885832e-05, "loss": 0.2748, "step": 112 }, { "epoch": 0.10482374768089053, "grad_norm": 3.676673173904419, "learning_rate": 1.9998688836656322e-05, "loss": 0.2271, "step": 113 }, { "epoch": 0.10575139146567718, "grad_norm": 4.632993698120117, "learning_rate": 1.9998111942939727e-05, "loss": 0.292, "step": 114 }, { "epoch": 0.10667903525046382, "grad_norm": 4.095834732055664, "learning_rate": 1.9997430173759876e-05, "loss": 0.2222, "step": 115 }, { "epoch": 0.10760667903525047, "grad_norm": 5.404327392578125, "learning_rate": 1.9996643536268202e-05, "loss": 0.3083, "step": 116 }, { "epoch": 0.10853432282003711, "grad_norm": 5.261657238006592, "learning_rate": 1.9995752038716166e-05, "loss": 0.2628, "step": 117 }, { "epoch": 0.10946196660482375, "grad_norm": 5.081439018249512, "learning_rate": 1.9994755690455154e-05, "loss": 0.3403, "step": 118 }, { "epoch": 0.11038961038961038, "grad_norm": 8.73869800567627, "learning_rate": 1.999365450193638e-05, "loss": 0.2949, "step": 119 }, { "epoch": 0.11131725417439703, "grad_norm": 6.3659586906433105, "learning_rate": 1.99924484847108e-05, "loss": 0.2925, "step": 120 }, { "epoch": 0.11131725417439703, "eval_accuracy": 0.8015521064301552, "eval_f1": 0.5095890410958904, "eval_loss": 0.500335693359375, "eval_precision": 0.8303571428571429, "eval_recall": 0.3675889328063241, "eval_runtime": 48.0846, "eval_samples_per_second": 5.74, "eval_steps_per_second": 0.187, "step": 120 }, { "epoch": 0.11224489795918367, "grad_norm": 4.58190393447876, "learning_rate": 1.9991137651428957e-05, "loss": 0.2265, "step": 121 }, { "epoch": 0.11317254174397032, "grad_norm": 5.173889636993408, "learning_rate": 1.998972201584088e-05, "loss": 0.2239, "step": 122 }, { "epoch": 0.11410018552875696, "grad_norm": 4.783614158630371, "learning_rate": 1.998820159279591e-05, "loss": 0.2998, "step": 123 }, { "epoch": 0.1150278293135436, "grad_norm": 3.5412709712982178, "learning_rate": 1.9986576398242566e-05, "loss": 0.2021, "step": 124 }, { "epoch": 0.11595547309833024, "grad_norm": 3.692047119140625, "learning_rate": 1.998484644922837e-05, "loss": 0.2432, "step": 125 }, { "epoch": 0.11688311688311688, "grad_norm": 2.976855993270874, "learning_rate": 1.9983011763899674e-05, "loss": 0.2703, "step": 126 }, { "epoch": 0.11781076066790352, "grad_norm": 5.132311820983887, "learning_rate": 1.998107236150145e-05, "loss": 0.3625, "step": 127 }, { "epoch": 0.11873840445269017, "grad_norm": 5.332205772399902, "learning_rate": 1.997902826237712e-05, "loss": 0.3431, "step": 128 }, { "epoch": 0.11966604823747681, "grad_norm": 7.54325532913208, "learning_rate": 1.997687948796831e-05, "loss": 0.2762, "step": 129 }, { "epoch": 0.12059369202226346, "grad_norm": 4.39344596862793, "learning_rate": 1.997462606081465e-05, "loss": 0.2178, "step": 130 }, { "epoch": 0.12152133580705009, "grad_norm": 4.67897891998291, "learning_rate": 1.997226800455352e-05, "loss": 0.2575, "step": 131 }, { "epoch": 0.12244897959183673, "grad_norm": 2.7064077854156494, "learning_rate": 1.9969805343919822e-05, "loss": 0.1973, "step": 132 }, { "epoch": 0.12337662337662338, "grad_norm": 2.841456413269043, "learning_rate": 1.9967238104745695e-05, "loss": 0.186, "step": 133 }, { "epoch": 0.12430426716141002, "grad_norm": 5.078066349029541, "learning_rate": 1.9964566313960265e-05, "loss": 0.2899, "step": 134 }, { "epoch": 0.12523191094619665, "grad_norm": 3.5166287422180176, "learning_rate": 1.9961789999589357e-05, "loss": 0.2397, "step": 135 }, { "epoch": 0.1261595547309833, "grad_norm": 3.0311009883880615, "learning_rate": 1.995890919075519e-05, "loss": 0.2042, "step": 136 }, { "epoch": 0.12708719851576994, "grad_norm": 4.455051898956299, "learning_rate": 1.995592391767608e-05, "loss": 0.1938, "step": 137 }, { "epoch": 0.1280148423005566, "grad_norm": 3.05238676071167, "learning_rate": 1.995283421166614e-05, "loss": 0.2134, "step": 138 }, { "epoch": 0.12894248608534323, "grad_norm": 2.5545527935028076, "learning_rate": 1.994964010513492e-05, "loss": 0.1558, "step": 139 }, { "epoch": 0.12987012987012986, "grad_norm": 3.169755458831787, "learning_rate": 1.9946341631587086e-05, "loss": 0.1912, "step": 140 }, { "epoch": 0.12987012987012986, "eval_accuracy": 0.8004434589800443, "eval_f1": 0.5, "eval_loss": 0.45746758580207825, "eval_precision": 0.8411214953271028, "eval_recall": 0.3557312252964427, "eval_runtime": 48.503, "eval_samples_per_second": 5.69, "eval_steps_per_second": 0.186, "step": 140 }, { "epoch": 0.13079777365491652, "grad_norm": 5.176037788391113, "learning_rate": 1.9942938825622064e-05, "loss": 0.261, "step": 141 }, { "epoch": 0.13172541743970315, "grad_norm": 4.5571513175964355, "learning_rate": 1.9939431722933678e-05, "loss": 0.1861, "step": 142 }, { "epoch": 0.1326530612244898, "grad_norm": 4.165744304656982, "learning_rate": 1.993582036030978e-05, "loss": 0.252, "step": 143 }, { "epoch": 0.13358070500927643, "grad_norm": 3.6084752082824707, "learning_rate": 1.9932104775631847e-05, "loss": 0.2091, "step": 144 }, { "epoch": 0.1345083487940631, "grad_norm": 5.7025837898254395, "learning_rate": 1.992828500787461e-05, "loss": 0.2875, "step": 145 }, { "epoch": 0.13543599257884972, "grad_norm": 3.954706907272339, "learning_rate": 1.9924361097105624e-05, "loss": 0.147, "step": 146 }, { "epoch": 0.13636363636363635, "grad_norm": 3.246682643890381, "learning_rate": 1.992033308448486e-05, "loss": 0.1406, "step": 147 }, { "epoch": 0.137291280148423, "grad_norm": 7.386576175689697, "learning_rate": 1.9916201012264255e-05, "loss": 0.2637, "step": 148 }, { "epoch": 0.13821892393320964, "grad_norm": 11.797701835632324, "learning_rate": 1.9911964923787295e-05, "loss": 0.2989, "step": 149 }, { "epoch": 0.1391465677179963, "grad_norm": 4.424801349639893, "learning_rate": 1.990762486348855e-05, "loss": 0.2059, "step": 150 }, { "epoch": 0.14007421150278293, "grad_norm": 6.5464582443237305, "learning_rate": 1.9903180876893195e-05, "loss": 0.3335, "step": 151 }, { "epoch": 0.14100185528756956, "grad_norm": 6.232185363769531, "learning_rate": 1.989863301061654e-05, "loss": 0.1652, "step": 152 }, { "epoch": 0.14192949907235622, "grad_norm": 7.7785162925720215, "learning_rate": 1.9893981312363563e-05, "loss": 0.3246, "step": 153 }, { "epoch": 0.14285714285714285, "grad_norm": 4.034485340118408, "learning_rate": 1.9889225830928365e-05, "loss": 0.1639, "step": 154 }, { "epoch": 0.1437847866419295, "grad_norm": 4.565614223480225, "learning_rate": 1.9884366616193707e-05, "loss": 0.2567, "step": 155 }, { "epoch": 0.14471243042671614, "grad_norm": 4.671913146972656, "learning_rate": 1.987940371913044e-05, "loss": 0.2956, "step": 156 }, { "epoch": 0.1456400742115028, "grad_norm": 4.866475582122803, "learning_rate": 1.987433719179702e-05, "loss": 0.1732, "step": 157 }, { "epoch": 0.14656771799628943, "grad_norm": 7.748964786529541, "learning_rate": 1.9869167087338908e-05, "loss": 0.3068, "step": 158 }, { "epoch": 0.14749536178107606, "grad_norm": 3.243945837020874, "learning_rate": 1.986389345998806e-05, "loss": 0.2172, "step": 159 }, { "epoch": 0.14842300556586271, "grad_norm": 2.541400671005249, "learning_rate": 1.9858516365062334e-05, "loss": 0.1991, "step": 160 }, { "epoch": 0.14842300556586271, "eval_accuracy": 0.811529933481153, "eval_f1": 0.5478723404255319, "eval_loss": 0.4108695983886719, "eval_precision": 0.8373983739837398, "eval_recall": 0.40711462450592883, "eval_runtime": 49.0091, "eval_samples_per_second": 5.632, "eval_steps_per_second": 0.184, "step": 160 }, { "epoch": 0.14935064935064934, "grad_norm": 3.5915067195892334, "learning_rate": 1.9853035858964907e-05, "loss": 0.2252, "step": 161 }, { "epoch": 0.150278293135436, "grad_norm": 3.297874927520752, "learning_rate": 1.9847451999183692e-05, "loss": 0.2025, "step": 162 }, { "epoch": 0.15120593692022263, "grad_norm": 7.80188512802124, "learning_rate": 1.9841764844290744e-05, "loss": 0.3563, "step": 163 }, { "epoch": 0.15213358070500926, "grad_norm": 4.962357044219971, "learning_rate": 1.9835974453941623e-05, "loss": 0.2331, "step": 164 }, { "epoch": 0.15306122448979592, "grad_norm": 4.794024467468262, "learning_rate": 1.983008088887478e-05, "loss": 0.2759, "step": 165 }, { "epoch": 0.15398886827458255, "grad_norm": 5.007259368896484, "learning_rate": 1.9824084210910924e-05, "loss": 0.1732, "step": 166 }, { "epoch": 0.1549165120593692, "grad_norm": 4.154080390930176, "learning_rate": 1.9817984482952378e-05, "loss": 0.199, "step": 167 }, { "epoch": 0.15584415584415584, "grad_norm": 5.933828830718994, "learning_rate": 1.9811781768982392e-05, "loss": 0.3237, "step": 168 }, { "epoch": 0.1567717996289425, "grad_norm": 4.5709943771362305, "learning_rate": 1.980547613406451e-05, "loss": 0.2356, "step": 169 }, { "epoch": 0.15769944341372913, "grad_norm": 4.232250690460205, "learning_rate": 1.9799067644341844e-05, "loss": 0.2318, "step": 170 }, { "epoch": 0.15862708719851576, "grad_norm": 5.491664886474609, "learning_rate": 1.9792556367036432e-05, "loss": 0.2551, "step": 171 }, { "epoch": 0.15955473098330242, "grad_norm": 3.584186315536499, "learning_rate": 1.978594237044849e-05, "loss": 0.1733, "step": 172 }, { "epoch": 0.16048237476808905, "grad_norm": 2.8872857093811035, "learning_rate": 1.977922572395571e-05, "loss": 0.2076, "step": 173 }, { "epoch": 0.1614100185528757, "grad_norm": 4.199950695037842, "learning_rate": 1.977240649801253e-05, "loss": 0.2076, "step": 174 }, { "epoch": 0.16233766233766234, "grad_norm": 3.1421799659729004, "learning_rate": 1.9765484764149413e-05, "loss": 0.2281, "step": 175 }, { "epoch": 0.16326530612244897, "grad_norm": 4.832352638244629, "learning_rate": 1.9758460594972068e-05, "loss": 0.1834, "step": 176 }, { "epoch": 0.16419294990723562, "grad_norm": 2.8222384452819824, "learning_rate": 1.9751334064160708e-05, "loss": 0.1908, "step": 177 }, { "epoch": 0.16512059369202226, "grad_norm": 3.1630570888519287, "learning_rate": 1.9744105246469264e-05, "loss": 0.1962, "step": 178 }, { "epoch": 0.1660482374768089, "grad_norm": 3.811518669128418, "learning_rate": 1.9736774217724614e-05, "loss": 0.1786, "step": 179 }, { "epoch": 0.16697588126159554, "grad_norm": 4.223273754119873, "learning_rate": 1.9729341054825783e-05, "loss": 0.2153, "step": 180 }, { "epoch": 0.16697588126159554, "eval_accuracy": 0.8337028824833703, "eval_f1": 0.6268656716417911, "eval_loss": 0.37183114886283875, "eval_precision": 0.8456375838926175, "eval_recall": 0.4980237154150198, "eval_runtime": 48.5694, "eval_samples_per_second": 5.683, "eval_steps_per_second": 0.185, "step": 180 }, { "epoch": 0.1679035250463822, "grad_norm": 3.9425668716430664, "learning_rate": 1.972180583574313e-05, "loss": 0.1998, "step": 181 }, { "epoch": 0.16883116883116883, "grad_norm": 5.5068840980529785, "learning_rate": 1.9714168639517543e-05, "loss": 0.2466, "step": 182 }, { "epoch": 0.16975881261595546, "grad_norm": 6.162604808807373, "learning_rate": 1.9706429546259592e-05, "loss": 0.163, "step": 183 }, { "epoch": 0.17068645640074212, "grad_norm": 5.026734828948975, "learning_rate": 1.9698588637148705e-05, "loss": 0.275, "step": 184 }, { "epoch": 0.17161410018552875, "grad_norm": 6.298387050628662, "learning_rate": 1.9690645994432307e-05, "loss": 0.1692, "step": 185 }, { "epoch": 0.1725417439703154, "grad_norm": 5.307831287384033, "learning_rate": 1.9682601701424958e-05, "loss": 0.2499, "step": 186 }, { "epoch": 0.17346938775510204, "grad_norm": 6.9988203048706055, "learning_rate": 1.9674455842507494e-05, "loss": 0.2434, "step": 187 }, { "epoch": 0.17439703153988867, "grad_norm": 4.128062725067139, "learning_rate": 1.9666208503126115e-05, "loss": 0.1976, "step": 188 }, { "epoch": 0.17532467532467533, "grad_norm": 3.3845396041870117, "learning_rate": 1.9657859769791506e-05, "loss": 0.1355, "step": 189 }, { "epoch": 0.17625231910946196, "grad_norm": 3.59576416015625, "learning_rate": 1.9649409730077934e-05, "loss": 0.2027, "step": 190 }, { "epoch": 0.17717996289424862, "grad_norm": 3.110616683959961, "learning_rate": 1.9640858472622316e-05, "loss": 0.2039, "step": 191 }, { "epoch": 0.17810760667903525, "grad_norm": 3.8708298206329346, "learning_rate": 1.9632206087123296e-05, "loss": 0.2163, "step": 192 }, { "epoch": 0.1790352504638219, "grad_norm": 5.087402820587158, "learning_rate": 1.9623452664340305e-05, "loss": 0.2631, "step": 193 }, { "epoch": 0.17996289424860853, "grad_norm": 3.7053322792053223, "learning_rate": 1.9614598296092603e-05, "loss": 0.2034, "step": 194 }, { "epoch": 0.18089053803339517, "grad_norm": 4.846376419067383, "learning_rate": 1.9605643075258323e-05, "loss": 0.2071, "step": 195 }, { "epoch": 0.18181818181818182, "grad_norm": 3.7219040393829346, "learning_rate": 1.9596587095773496e-05, "loss": 0.2516, "step": 196 }, { "epoch": 0.18274582560296845, "grad_norm": 5.604904651641846, "learning_rate": 1.958743045263106e-05, "loss": 0.2076, "step": 197 }, { "epoch": 0.1836734693877551, "grad_norm": 2.9652745723724365, "learning_rate": 1.957817324187987e-05, "loss": 0.1752, "step": 198 }, { "epoch": 0.18460111317254174, "grad_norm": 4.468489646911621, "learning_rate": 1.956881556062369e-05, "loss": 0.2177, "step": 199 }, { "epoch": 0.18552875695732837, "grad_norm": 4.231448173522949, "learning_rate": 1.9559357507020163e-05, "loss": 0.1638, "step": 200 }, { "epoch": 0.18552875695732837, "eval_accuracy": 0.8237250554323725, "eval_f1": 0.5826771653543307, "eval_loss": 0.36573752760887146, "eval_precision": 0.8671875, "eval_recall": 0.43873517786561267, "eval_runtime": 49.5575, "eval_samples_per_second": 5.569, "eval_steps_per_second": 0.182, "step": 200 }, { "epoch": 0.18645640074211503, "grad_norm": 3.4353301525115967, "learning_rate": 1.9549799180279793e-05, "loss": 0.2293, "step": 201 }, { "epoch": 0.18738404452690166, "grad_norm": 3.4561715126037598, "learning_rate": 1.9540140680664915e-05, "loss": 0.1131, "step": 202 }, { "epoch": 0.18831168831168832, "grad_norm": 4.9482293128967285, "learning_rate": 1.953038210948861e-05, "loss": 0.2009, "step": 203 }, { "epoch": 0.18923933209647495, "grad_norm": 4.263972759246826, "learning_rate": 1.952052356911368e-05, "loss": 0.2624, "step": 204 }, { "epoch": 0.1901669758812616, "grad_norm": 4.766571044921875, "learning_rate": 1.9510565162951538e-05, "loss": 0.1823, "step": 205 }, { "epoch": 0.19109461966604824, "grad_norm": 5.494351387023926, "learning_rate": 1.950050699546116e-05, "loss": 0.2365, "step": 206 }, { "epoch": 0.19202226345083487, "grad_norm": 5.0484795570373535, "learning_rate": 1.9490349172147964e-05, "loss": 0.2197, "step": 207 }, { "epoch": 0.19294990723562153, "grad_norm": 4.198167324066162, "learning_rate": 1.9480091799562706e-05, "loss": 0.1825, "step": 208 }, { "epoch": 0.19387755102040816, "grad_norm": 4.7388105392456055, "learning_rate": 1.9469734985300373e-05, "loss": 0.2195, "step": 209 }, { "epoch": 0.19480519480519481, "grad_norm": 4.556212902069092, "learning_rate": 1.9459278837999048e-05, "loss": 0.2085, "step": 210 }, { "epoch": 0.19573283858998144, "grad_norm": 5.153113842010498, "learning_rate": 1.9448723467338765e-05, "loss": 0.2332, "step": 211 }, { "epoch": 0.19666048237476808, "grad_norm": 2.6720712184906006, "learning_rate": 1.9438068984040366e-05, "loss": 0.1761, "step": 212 }, { "epoch": 0.19758812615955473, "grad_norm": 2.9613466262817383, "learning_rate": 1.9427315499864345e-05, "loss": 0.1877, "step": 213 }, { "epoch": 0.19851576994434136, "grad_norm": 3.5731585025787354, "learning_rate": 1.9416463127609655e-05, "loss": 0.2018, "step": 214 }, { "epoch": 0.19944341372912802, "grad_norm": 4.205770015716553, "learning_rate": 1.9405511981112553e-05, "loss": 0.1952, "step": 215 }, { "epoch": 0.20037105751391465, "grad_norm": 4.905291557312012, "learning_rate": 1.9394462175245382e-05, "loss": 0.2482, "step": 216 }, { "epoch": 0.2012987012987013, "grad_norm": 6.575018405914307, "learning_rate": 1.9383313825915372e-05, "loss": 0.2007, "step": 217 }, { "epoch": 0.20222634508348794, "grad_norm": 4.795684814453125, "learning_rate": 1.937206705006344e-05, "loss": 0.1878, "step": 218 }, { "epoch": 0.20315398886827457, "grad_norm": 3.1328468322753906, "learning_rate": 1.9360721965662934e-05, "loss": 0.1864, "step": 219 }, { "epoch": 0.20408163265306123, "grad_norm": 3.297593355178833, "learning_rate": 1.9349278691718426e-05, "loss": 0.2033, "step": 220 }, { "epoch": 0.20408163265306123, "eval_accuracy": 0.8370288248337029, "eval_f1": 0.6423357664233577, "eval_loss": 0.34552833437919617, "eval_precision": 0.8354430379746836, "eval_recall": 0.5217391304347826, "eval_runtime": 48.9715, "eval_samples_per_second": 5.636, "eval_steps_per_second": 0.184, "step": 220 }, { "epoch": 0.20500927643784786, "grad_norm": 3.253046989440918, "learning_rate": 1.9337737348264448e-05, "loss": 0.2099, "step": 221 }, { "epoch": 0.20593692022263452, "grad_norm": 4.903271675109863, "learning_rate": 1.9326098056364224e-05, "loss": 0.2595, "step": 222 }, { "epoch": 0.20686456400742115, "grad_norm": 6.408766746520996, "learning_rate": 1.9314360938108427e-05, "loss": 0.2615, "step": 223 }, { "epoch": 0.2077922077922078, "grad_norm": 4.407992839813232, "learning_rate": 1.9302526116613863e-05, "loss": 0.1009, "step": 224 }, { "epoch": 0.20871985157699444, "grad_norm": 2.9286835193634033, "learning_rate": 1.9290593716022218e-05, "loss": 0.1412, "step": 225 }, { "epoch": 0.20964749536178107, "grad_norm": 3.6685068607330322, "learning_rate": 1.9278563861498726e-05, "loss": 0.221, "step": 226 }, { "epoch": 0.21057513914656772, "grad_norm": 4.300150394439697, "learning_rate": 1.9266436679230866e-05, "loss": 0.1645, "step": 227 }, { "epoch": 0.21150278293135436, "grad_norm": 4.307882308959961, "learning_rate": 1.9254212296427043e-05, "loss": 0.2284, "step": 228 }, { "epoch": 0.212430426716141, "grad_norm": 2.4391989707946777, "learning_rate": 1.924189084131525e-05, "loss": 0.118, "step": 229 }, { "epoch": 0.21335807050092764, "grad_norm": 5.997758388519287, "learning_rate": 1.922947244314172e-05, "loss": 0.2798, "step": 230 }, { "epoch": 0.21428571428571427, "grad_norm": 4.957704544067383, "learning_rate": 1.9216957232169567e-05, "loss": 0.1852, "step": 231 }, { "epoch": 0.21521335807050093, "grad_norm": 5.667599678039551, "learning_rate": 1.9204345339677442e-05, "loss": 0.2155, "step": 232 }, { "epoch": 0.21614100185528756, "grad_norm": 4.291685104370117, "learning_rate": 1.9191636897958123e-05, "loss": 0.2637, "step": 233 }, { "epoch": 0.21706864564007422, "grad_norm": 4.587561130523682, "learning_rate": 1.9178832040317153e-05, "loss": 0.2383, "step": 234 }, { "epoch": 0.21799628942486085, "grad_norm": 3.8426513671875, "learning_rate": 1.916593090107143e-05, "loss": 0.212, "step": 235 }, { "epoch": 0.2189239332096475, "grad_norm": 5.173671722412109, "learning_rate": 1.91529336155478e-05, "loss": 0.2054, "step": 236 }, { "epoch": 0.21985157699443414, "grad_norm": 4.827680587768555, "learning_rate": 1.913984032008163e-05, "loss": 0.2072, "step": 237 }, { "epoch": 0.22077922077922077, "grad_norm": 7.693399906158447, "learning_rate": 1.9126651152015404e-05, "loss": 0.2071, "step": 238 }, { "epoch": 0.22170686456400743, "grad_norm": 3.128953695297241, "learning_rate": 1.911336624969725e-05, "loss": 0.2271, "step": 239 }, { "epoch": 0.22263450834879406, "grad_norm": 6.966447353363037, "learning_rate": 1.9099985752479505e-05, "loss": 0.2448, "step": 240 }, { "epoch": 0.22263450834879406, "eval_accuracy": 0.8381374722838137, "eval_f1": 0.6403940886699507, "eval_loss": 0.34381967782974243, "eval_precision": 0.8496732026143791, "eval_recall": 0.5138339920948617, "eval_runtime": 48.6504, "eval_samples_per_second": 5.673, "eval_steps_per_second": 0.185, "step": 240 }, { "epoch": 0.22356215213358072, "grad_norm": 4.817497730255127, "learning_rate": 1.908650980071726e-05, "loss": 0.1878, "step": 241 }, { "epoch": 0.22448979591836735, "grad_norm": 3.612339973449707, "learning_rate": 1.9072938535766864e-05, "loss": 0.1226, "step": 242 }, { "epoch": 0.22541743970315398, "grad_norm": 4.005315780639648, "learning_rate": 1.905927209998447e-05, "loss": 0.2076, "step": 243 }, { "epoch": 0.22634508348794063, "grad_norm": 3.032423496246338, "learning_rate": 1.904551063672452e-05, "loss": 0.1511, "step": 244 }, { "epoch": 0.22727272727272727, "grad_norm": 3.060365676879883, "learning_rate": 1.9031654290338256e-05, "loss": 0.1954, "step": 245 }, { "epoch": 0.22820037105751392, "grad_norm": 5.271503448486328, "learning_rate": 1.9017703206172187e-05, "loss": 0.2244, "step": 246 }, { "epoch": 0.22912801484230055, "grad_norm": 2.923628568649292, "learning_rate": 1.900365753056659e-05, "loss": 0.1753, "step": 247 }, { "epoch": 0.2300556586270872, "grad_norm": 3.745664119720459, "learning_rate": 1.8989517410853956e-05, "loss": 0.188, "step": 248 }, { "epoch": 0.23098330241187384, "grad_norm": 3.7633256912231445, "learning_rate": 1.8975282995357448e-05, "loss": 0.201, "step": 249 }, { "epoch": 0.23191094619666047, "grad_norm": 3.29656720161438, "learning_rate": 1.896095443338935e-05, "loss": 0.1956, "step": 250 }, { "epoch": 0.23283858998144713, "grad_norm": 4.702951431274414, "learning_rate": 1.8946531875249496e-05, "loss": 0.2101, "step": 251 }, { "epoch": 0.23376623376623376, "grad_norm": 7.016535758972168, "learning_rate": 1.8932015472223692e-05, "loss": 0.2486, "step": 252 }, { "epoch": 0.23469387755102042, "grad_norm": 4.004672050476074, "learning_rate": 1.8917405376582144e-05, "loss": 0.239, "step": 253 }, { "epoch": 0.23562152133580705, "grad_norm": 4.382692337036133, "learning_rate": 1.8902701741577844e-05, "loss": 0.2308, "step": 254 }, { "epoch": 0.23654916512059368, "grad_norm": 4.72487735748291, "learning_rate": 1.8887904721444955e-05, "loss": 0.2098, "step": 255 }, { "epoch": 0.23747680890538034, "grad_norm": 3.4035651683807373, "learning_rate": 1.8873014471397225e-05, "loss": 0.101, "step": 256 }, { "epoch": 0.23840445269016697, "grad_norm": 7.383582592010498, "learning_rate": 1.8858031147626326e-05, "loss": 0.2499, "step": 257 }, { "epoch": 0.23933209647495363, "grad_norm": 4.3388895988464355, "learning_rate": 1.8842954907300236e-05, "loss": 0.1859, "step": 258 }, { "epoch": 0.24025974025974026, "grad_norm": 3.2572548389434814, "learning_rate": 1.8827785908561585e-05, "loss": 0.1833, "step": 259 }, { "epoch": 0.24118738404452691, "grad_norm": 7.2369561195373535, "learning_rate": 1.881252431052599e-05, "loss": 0.2337, "step": 260 }, { "epoch": 0.24118738404452691, "eval_accuracy": 0.8281596452328159, "eval_f1": 0.5931758530183727, "eval_loss": 0.3704891502857208, "eval_precision": 0.8828125, "eval_recall": 0.44664031620553357, "eval_runtime": 47.7821, "eval_samples_per_second": 5.776, "eval_steps_per_second": 0.188, "step": 260 }, { "epoch": 0.24211502782931354, "grad_norm": 6.052790641784668, "learning_rate": 1.879717027328039e-05, "loss": 0.2569, "step": 261 }, { "epoch": 0.24304267161410018, "grad_norm": 3.246156692504883, "learning_rate": 1.8781723957881374e-05, "loss": 0.1871, "step": 262 }, { "epoch": 0.24397031539888683, "grad_norm": 6.536667823791504, "learning_rate": 1.876618552635348e-05, "loss": 0.1425, "step": 263 }, { "epoch": 0.24489795918367346, "grad_norm": 5.4920830726623535, "learning_rate": 1.87505551416875e-05, "loss": 0.2275, "step": 264 }, { "epoch": 0.24582560296846012, "grad_norm": 7.879229545593262, "learning_rate": 1.8734832967838775e-05, "loss": 0.2852, "step": 265 }, { "epoch": 0.24675324675324675, "grad_norm": 3.7755329608917236, "learning_rate": 1.871901916972547e-05, "loss": 0.1967, "step": 266 }, { "epoch": 0.24768089053803338, "grad_norm": 4.236266136169434, "learning_rate": 1.8703113913226847e-05, "loss": 0.1302, "step": 267 }, { "epoch": 0.24860853432282004, "grad_norm": 5.8521599769592285, "learning_rate": 1.8687117365181514e-05, "loss": 0.2866, "step": 268 }, { "epoch": 0.24953617810760667, "grad_norm": 2.9448065757751465, "learning_rate": 1.867102969338569e-05, "loss": 0.171, "step": 269 }, { "epoch": 0.2504638218923933, "grad_norm": 4.411244869232178, "learning_rate": 1.865485106659145e-05, "loss": 0.1424, "step": 270 }, { "epoch": 0.25139146567717996, "grad_norm": 5.2516655921936035, "learning_rate": 1.863858165450492e-05, "loss": 0.2166, "step": 271 }, { "epoch": 0.2523191094619666, "grad_norm": 2.6857924461364746, "learning_rate": 1.862222162778454e-05, "loss": 0.1786, "step": 272 }, { "epoch": 0.2532467532467532, "grad_norm": 6.46083927154541, "learning_rate": 1.8605771158039253e-05, "loss": 0.1817, "step": 273 }, { "epoch": 0.2541743970315399, "grad_norm": 6.851802349090576, "learning_rate": 1.85892304178267e-05, "loss": 0.174, "step": 274 }, { "epoch": 0.25510204081632654, "grad_norm": 4.372910022735596, "learning_rate": 1.8572599580651415e-05, "loss": 0.211, "step": 275 }, { "epoch": 0.2560296846011132, "grad_norm": 8.214804649353027, "learning_rate": 1.8555878820963014e-05, "loss": 0.2295, "step": 276 }, { "epoch": 0.2569573283858998, "grad_norm": 3.299901247024536, "learning_rate": 1.8539068314154355e-05, "loss": 0.1784, "step": 277 }, { "epoch": 0.25788497217068646, "grad_norm": 3.1576666831970215, "learning_rate": 1.8522168236559693e-05, "loss": 0.1864, "step": 278 }, { "epoch": 0.2588126159554731, "grad_norm": 6.918071269989014, "learning_rate": 1.8505178765452853e-05, "loss": 0.2015, "step": 279 }, { "epoch": 0.2597402597402597, "grad_norm": 2.884896755218506, "learning_rate": 1.8488100079045345e-05, "loss": 0.1698, "step": 280 }, { "epoch": 0.2597402597402597, "eval_accuracy": 0.8215077605321508, "eval_f1": 0.5729442970822282, "eval_loss": 0.37236273288726807, "eval_precision": 0.8709677419354839, "eval_recall": 0.4268774703557312, "eval_runtime": 48.19, "eval_samples_per_second": 5.727, "eval_steps_per_second": 0.187, "step": 280 }, { "epoch": 0.2606679035250464, "grad_norm": 2.52854061126709, "learning_rate": 1.847093235648451e-05, "loss": 0.1427, "step": 281 }, { "epoch": 0.26159554730983303, "grad_norm": 3.8434646129608154, "learning_rate": 1.8453675777851627e-05, "loss": 0.2429, "step": 282 }, { "epoch": 0.2625231910946197, "grad_norm": 3.32183837890625, "learning_rate": 1.8436330524160048e-05, "loss": 0.1672, "step": 283 }, { "epoch": 0.2634508348794063, "grad_norm": 3.077954053878784, "learning_rate": 1.8418896777353272e-05, "loss": 0.1511, "step": 284 }, { "epoch": 0.26437847866419295, "grad_norm": 6.1399617195129395, "learning_rate": 1.8401374720303054e-05, "loss": 0.1989, "step": 285 }, { "epoch": 0.2653061224489796, "grad_norm": 4.476570129394531, "learning_rate": 1.8383764536807486e-05, "loss": 0.2204, "step": 286 }, { "epoch": 0.2662337662337662, "grad_norm": 3.3272931575775146, "learning_rate": 1.836606641158905e-05, "loss": 0.1733, "step": 287 }, { "epoch": 0.26716141001855287, "grad_norm": 3.8696646690368652, "learning_rate": 1.8348280530292712e-05, "loss": 0.2698, "step": 288 }, { "epoch": 0.2680890538033395, "grad_norm": 4.347325325012207, "learning_rate": 1.833040707948395e-05, "loss": 0.2098, "step": 289 }, { "epoch": 0.2690166975881262, "grad_norm": 3.8024332523345947, "learning_rate": 1.831244624664681e-05, "loss": 0.2092, "step": 290 }, { "epoch": 0.2699443413729128, "grad_norm": 3.742701768875122, "learning_rate": 1.829439822018192e-05, "loss": 0.1793, "step": 291 }, { "epoch": 0.27087198515769945, "grad_norm": 4.459840774536133, "learning_rate": 1.827626318940454e-05, "loss": 0.224, "step": 292 }, { "epoch": 0.2717996289424861, "grad_norm": 3.4559335708618164, "learning_rate": 1.8258041344542567e-05, "loss": 0.2342, "step": 293 }, { "epoch": 0.2727272727272727, "grad_norm": 4.520707607269287, "learning_rate": 1.8239732876734525e-05, "loss": 0.2899, "step": 294 }, { "epoch": 0.27365491651205937, "grad_norm": 3.844388246536255, "learning_rate": 1.822133797802758e-05, "loss": 0.172, "step": 295 }, { "epoch": 0.274582560296846, "grad_norm": 3.944843292236328, "learning_rate": 1.8202856841375517e-05, "loss": 0.1602, "step": 296 }, { "epoch": 0.2755102040816326, "grad_norm": 2.833136796951294, "learning_rate": 1.8184289660636715e-05, "loss": 0.1829, "step": 297 }, { "epoch": 0.2764378478664193, "grad_norm": 5.877793312072754, "learning_rate": 1.816563663057211e-05, "loss": 0.1776, "step": 298 }, { "epoch": 0.27736549165120594, "grad_norm": 7.863223552703857, "learning_rate": 1.8146897946843162e-05, "loss": 0.2734, "step": 299 }, { "epoch": 0.2782931354359926, "grad_norm": 3.5865793228149414, "learning_rate": 1.81280738060098e-05, "loss": 0.1607, "step": 300 }, { "epoch": 0.2782931354359926, "eval_accuracy": 0.8292682926829268, "eval_f1": 0.6010362694300518, "eval_loss": 0.34550240635871887, "eval_precision": 0.8721804511278195, "eval_recall": 0.45849802371541504, "eval_runtime": 47.9887, "eval_samples_per_second": 5.751, "eval_steps_per_second": 0.188, "step": 300 }, { "epoch": 0.2792207792207792, "grad_norm": 4.416810512542725, "learning_rate": 1.810916440552835e-05, "loss": 0.2122, "step": 301 }, { "epoch": 0.28014842300556586, "grad_norm": 4.608455657958984, "learning_rate": 1.8090169943749477e-05, "loss": 0.1634, "step": 302 }, { "epoch": 0.2810760667903525, "grad_norm": 2.9112977981567383, "learning_rate": 1.8071090619916095e-05, "loss": 0.1863, "step": 303 }, { "epoch": 0.2820037105751391, "grad_norm": 4.165685176849365, "learning_rate": 1.8051926634161282e-05, "loss": 0.1607, "step": 304 }, { "epoch": 0.2829313543599258, "grad_norm": 3.188124895095825, "learning_rate": 1.8032678187506187e-05, "loss": 0.1379, "step": 305 }, { "epoch": 0.28385899814471244, "grad_norm": 4.277055740356445, "learning_rate": 1.8013345481857903e-05, "loss": 0.1945, "step": 306 }, { "epoch": 0.2847866419294991, "grad_norm": 4.259377956390381, "learning_rate": 1.799392872000736e-05, "loss": 0.2123, "step": 307 }, { "epoch": 0.2857142857142857, "grad_norm": 4.486137390136719, "learning_rate": 1.797442810562721e-05, "loss": 0.2004, "step": 308 }, { "epoch": 0.28664192949907236, "grad_norm": 6.125461578369141, "learning_rate": 1.7954843843269665e-05, "loss": 0.2124, "step": 309 }, { "epoch": 0.287569573283859, "grad_norm": 4.838400363922119, "learning_rate": 1.793517613836437e-05, "loss": 0.2112, "step": 310 }, { "epoch": 0.2884972170686456, "grad_norm": 3.1725666522979736, "learning_rate": 1.7915425197216246e-05, "loss": 0.1658, "step": 311 }, { "epoch": 0.2894248608534323, "grad_norm": 3.154855489730835, "learning_rate": 1.7895591227003316e-05, "loss": 0.2081, "step": 312 }, { "epoch": 0.29035250463821893, "grad_norm": 3.7664122581481934, "learning_rate": 1.7875674435774546e-05, "loss": 0.1628, "step": 313 }, { "epoch": 0.2912801484230056, "grad_norm": 5.876150131225586, "learning_rate": 1.7855675032447648e-05, "loss": 0.1882, "step": 314 }, { "epoch": 0.2922077922077922, "grad_norm": 2.4536941051483154, "learning_rate": 1.7835593226806902e-05, "loss": 0.1287, "step": 315 }, { "epoch": 0.29313543599257885, "grad_norm": 4.552228927612305, "learning_rate": 1.7815429229500946e-05, "loss": 0.2509, "step": 316 }, { "epoch": 0.2940630797773655, "grad_norm": 3.861381769180298, "learning_rate": 1.7795183252040568e-05, "loss": 0.1827, "step": 317 }, { "epoch": 0.2949907235621521, "grad_norm": 3.6626925468444824, "learning_rate": 1.7774855506796497e-05, "loss": 0.154, "step": 318 }, { "epoch": 0.29591836734693877, "grad_norm": 6.695428848266602, "learning_rate": 1.7754446206997152e-05, "loss": 0.2334, "step": 319 }, { "epoch": 0.29684601113172543, "grad_norm": 3.2786192893981934, "learning_rate": 1.7733955566726438e-05, "loss": 0.1671, "step": 320 }, { "epoch": 0.29684601113172543, "eval_accuracy": 0.8337028824833703, "eval_f1": 0.625, "eval_loss": 0.3371331989765167, "eval_precision": 0.8503401360544217, "eval_recall": 0.49407114624505927, "eval_runtime": 47.664, "eval_samples_per_second": 5.791, "eval_steps_per_second": 0.189, "step": 320 }, { "epoch": 0.29777365491651203, "grad_norm": 4.181811809539795, "learning_rate": 1.771338380092148e-05, "loss": 0.2198, "step": 321 }, { "epoch": 0.2987012987012987, "grad_norm": 4.808495998382568, "learning_rate": 1.7692731125370355e-05, "loss": 0.171, "step": 322 }, { "epoch": 0.29962894248608535, "grad_norm": 5.946144104003906, "learning_rate": 1.767199775670986e-05, "loss": 0.2176, "step": 323 }, { "epoch": 0.300556586270872, "grad_norm": 4.010025978088379, "learning_rate": 1.7651183912423228e-05, "loss": 0.2003, "step": 324 }, { "epoch": 0.3014842300556586, "grad_norm": 4.795968055725098, "learning_rate": 1.7630289810837836e-05, "loss": 0.165, "step": 325 }, { "epoch": 0.30241187384044527, "grad_norm": 2.9344944953918457, "learning_rate": 1.7609315671122912e-05, "loss": 0.1603, "step": 326 }, { "epoch": 0.3033395176252319, "grad_norm": 5.812965393066406, "learning_rate": 1.758826171328727e-05, "loss": 0.1458, "step": 327 }, { "epoch": 0.3042671614100185, "grad_norm": 4.66487979888916, "learning_rate": 1.7567128158176955e-05, "loss": 0.2155, "step": 328 }, { "epoch": 0.3051948051948052, "grad_norm": 3.436579942703247, "learning_rate": 1.7545915227472967e-05, "loss": 0.1699, "step": 329 }, { "epoch": 0.30612244897959184, "grad_norm": 4.022927284240723, "learning_rate": 1.7524623143688905e-05, "loss": 0.2037, "step": 330 }, { "epoch": 0.3070500927643785, "grad_norm": 3.154623031616211, "learning_rate": 1.7503252130168657e-05, "loss": 0.1627, "step": 331 }, { "epoch": 0.3079777365491651, "grad_norm": 2.7966222763061523, "learning_rate": 1.748180241108404e-05, "loss": 0.1439, "step": 332 }, { "epoch": 0.30890538033395176, "grad_norm": 3.8477606773376465, "learning_rate": 1.7460274211432463e-05, "loss": 0.135, "step": 333 }, { "epoch": 0.3098330241187384, "grad_norm": 3.51237416267395, "learning_rate": 1.7438667757034547e-05, "loss": 0.213, "step": 334 }, { "epoch": 0.310760667903525, "grad_norm": 4.482552528381348, "learning_rate": 1.7416983274531777e-05, "loss": 0.2018, "step": 335 }, { "epoch": 0.3116883116883117, "grad_norm": 3.7399659156799316, "learning_rate": 1.739522099138411e-05, "loss": 0.2196, "step": 336 }, { "epoch": 0.31261595547309834, "grad_norm": 3.4632811546325684, "learning_rate": 1.7373381135867605e-05, "loss": 0.1588, "step": 337 }, { "epoch": 0.313543599257885, "grad_norm": 5.400178909301758, "learning_rate": 1.7351463937072008e-05, "loss": 0.2455, "step": 338 }, { "epoch": 0.3144712430426716, "grad_norm": 4.754225730895996, "learning_rate": 1.732946962489836e-05, "loss": 0.2067, "step": 339 }, { "epoch": 0.31539888682745826, "grad_norm": 3.6749982833862305, "learning_rate": 1.7307398430056595e-05, "loss": 0.1809, "step": 340 }, { "epoch": 0.31539888682745826, "eval_accuracy": 0.8514412416851441, "eval_f1": 0.6912442396313364, "eval_loss": 0.34059056639671326, "eval_precision": 0.8287292817679558, "eval_recall": 0.5928853754940712, "eval_runtime": 47.764, "eval_samples_per_second": 5.778, "eval_steps_per_second": 0.188, "step": 340 }, { "epoch": 0.3163265306122449, "grad_norm": 5.549361705780029, "learning_rate": 1.72852505840631e-05, "loss": 0.1656, "step": 341 }, { "epoch": 0.3172541743970315, "grad_norm": 7.491882801055908, "learning_rate": 1.72630263192383e-05, "loss": 0.2493, "step": 342 }, { "epoch": 0.3181818181818182, "grad_norm": 5.225057125091553, "learning_rate": 1.7240725868704218e-05, "loss": 0.2031, "step": 343 }, { "epoch": 0.31910946196660483, "grad_norm": 7.436551094055176, "learning_rate": 1.7218349466382024e-05, "loss": 0.2245, "step": 344 }, { "epoch": 0.3200371057513915, "grad_norm": 4.4297871589660645, "learning_rate": 1.719589734698959e-05, "loss": 0.2226, "step": 345 }, { "epoch": 0.3209647495361781, "grad_norm": 6.560292720794678, "learning_rate": 1.7173369746039026e-05, "loss": 0.2331, "step": 346 }, { "epoch": 0.32189239332096475, "grad_norm": 5.045467376708984, "learning_rate": 1.7150766899834205e-05, "loss": 0.2321, "step": 347 }, { "epoch": 0.3228200371057514, "grad_norm": 6.464671611785889, "learning_rate": 1.7128089045468294e-05, "loss": 0.19, "step": 348 }, { "epoch": 0.323747680890538, "grad_norm": 2.7538771629333496, "learning_rate": 1.7105336420821247e-05, "loss": 0.1642, "step": 349 }, { "epoch": 0.3246753246753247, "grad_norm": 4.287221431732178, "learning_rate": 1.7082509264557333e-05, "loss": 0.1556, "step": 350 }, { "epoch": 0.32560296846011133, "grad_norm": 3.1308867931365967, "learning_rate": 1.705960781612262e-05, "loss": 0.1583, "step": 351 }, { "epoch": 0.32653061224489793, "grad_norm": 3.598468542098999, "learning_rate": 1.7036632315742464e-05, "loss": 0.1571, "step": 352 }, { "epoch": 0.3274582560296846, "grad_norm": 2.93074369430542, "learning_rate": 1.7013583004418994e-05, "loss": 0.1463, "step": 353 }, { "epoch": 0.32838589981447125, "grad_norm": 3.5115561485290527, "learning_rate": 1.6990460123928577e-05, "loss": 0.1522, "step": 354 }, { "epoch": 0.3293135435992579, "grad_norm": 2.3975048065185547, "learning_rate": 1.696726391681929e-05, "loss": 0.136, "step": 355 }, { "epoch": 0.3302411873840445, "grad_norm": 3.01874041557312, "learning_rate": 1.6943994626408365e-05, "loss": 0.1928, "step": 356 }, { "epoch": 0.33116883116883117, "grad_norm": 3.620572805404663, "learning_rate": 1.692065249677965e-05, "loss": 0.1964, "step": 357 }, { "epoch": 0.3320964749536178, "grad_norm": 3.1630985736846924, "learning_rate": 1.6897237772781046e-05, "loss": 0.1757, "step": 358 }, { "epoch": 0.33302411873840443, "grad_norm": 2.7404632568359375, "learning_rate": 1.6873750700021917e-05, "loss": 0.1391, "step": 359 }, { "epoch": 0.3339517625231911, "grad_norm": 3.915576696395874, "learning_rate": 1.6850191524870548e-05, "loss": 0.1672, "step": 360 }, { "epoch": 0.3339517625231911, "eval_accuracy": 0.8392461197339246, "eval_f1": 0.6365914786967418, "eval_loss": 0.3519636392593384, "eval_precision": 0.8698630136986302, "eval_recall": 0.5019762845849802, "eval_runtime": 48.8319, "eval_samples_per_second": 5.652, "eval_steps_per_second": 0.184, "step": 360 }, { "epoch": 0.33487940630797774, "grad_norm": 4.442441940307617, "learning_rate": 1.682656049445154e-05, "loss": 0.1695, "step": 361 }, { "epoch": 0.3358070500927644, "grad_norm": 3.9514217376708984, "learning_rate": 1.6802857856643214e-05, "loss": 0.1744, "step": 362 }, { "epoch": 0.336734693877551, "grad_norm": 5.170446395874023, "learning_rate": 1.6779083860075032e-05, "loss": 0.1198, "step": 363 }, { "epoch": 0.33766233766233766, "grad_norm": 7.009990215301514, "learning_rate": 1.6755238754124965e-05, "loss": 0.1708, "step": 364 }, { "epoch": 0.3385899814471243, "grad_norm": 7.845787048339844, "learning_rate": 1.6731322788916892e-05, "loss": 0.1723, "step": 365 }, { "epoch": 0.3395176252319109, "grad_norm": 3.684112071990967, "learning_rate": 1.6707336215317968e-05, "loss": 0.1136, "step": 366 }, { "epoch": 0.3404452690166976, "grad_norm": 6.165255069732666, "learning_rate": 1.6683279284936004e-05, "loss": 0.2895, "step": 367 }, { "epoch": 0.34137291280148424, "grad_norm": 5.1027326583862305, "learning_rate": 1.665915225011681e-05, "loss": 0.1915, "step": 368 }, { "epoch": 0.3423005565862709, "grad_norm": 4.702267646789551, "learning_rate": 1.6634955363941573e-05, "loss": 0.135, "step": 369 }, { "epoch": 0.3432282003710575, "grad_norm": 3.623323678970337, "learning_rate": 1.6610688880224178e-05, "loss": 0.1364, "step": 370 }, { "epoch": 0.34415584415584416, "grad_norm": 4.77068567276001, "learning_rate": 1.6586353053508548e-05, "loss": 0.2082, "step": 371 }, { "epoch": 0.3450834879406308, "grad_norm": 6.354672908782959, "learning_rate": 1.6561948139065997e-05, "loss": 0.2037, "step": 372 }, { "epoch": 0.3460111317254174, "grad_norm": 3.074613094329834, "learning_rate": 1.6537474392892527e-05, "loss": 0.1123, "step": 373 }, { "epoch": 0.3469387755102041, "grad_norm": 5.388340473175049, "learning_rate": 1.6512932071706153e-05, "loss": 0.2308, "step": 374 }, { "epoch": 0.34786641929499074, "grad_norm": 3.377591609954834, "learning_rate": 1.6488321432944218e-05, "loss": 0.1607, "step": 375 }, { "epoch": 0.34879406307977734, "grad_norm": 3.697253465652466, "learning_rate": 1.646364273476067e-05, "loss": 0.2064, "step": 376 }, { "epoch": 0.349721706864564, "grad_norm": 2.825587272644043, "learning_rate": 1.6438896236023374e-05, "loss": 0.1405, "step": 377 }, { "epoch": 0.35064935064935066, "grad_norm": 5.332752227783203, "learning_rate": 1.6414082196311402e-05, "loss": 0.1838, "step": 378 }, { "epoch": 0.3515769944341373, "grad_norm": 3.3014256954193115, "learning_rate": 1.638920087591228e-05, "loss": 0.1516, "step": 379 }, { "epoch": 0.3525046382189239, "grad_norm": 2.9438254833221436, "learning_rate": 1.6364252535819284e-05, "loss": 0.153, "step": 380 }, { "epoch": 0.3525046382189239, "eval_accuracy": 0.8458980044345898, "eval_f1": 0.6634382566585957, "eval_loss": 0.32734009623527527, "eval_precision": 0.85625, "eval_recall": 0.541501976284585, "eval_runtime": 47.3489, "eval_samples_per_second": 5.829, "eval_steps_per_second": 0.19, "step": 380 }, { "epoch": 0.3534322820037106, "grad_norm": 4.24219274520874, "learning_rate": 1.63392374377287e-05, "loss": 0.1265, "step": 381 }, { "epoch": 0.35435992578849723, "grad_norm": 3.8825278282165527, "learning_rate": 1.6314155844037074e-05, "loss": 0.173, "step": 382 }, { "epoch": 0.35528756957328383, "grad_norm": 4.059886455535889, "learning_rate": 1.6289008017838447e-05, "loss": 0.1639, "step": 383 }, { "epoch": 0.3562152133580705, "grad_norm": 5.13779878616333, "learning_rate": 1.626379422292162e-05, "loss": 0.1006, "step": 384 }, { "epoch": 0.35714285714285715, "grad_norm": 5.269052028656006, "learning_rate": 1.6238514723767372e-05, "loss": 0.2033, "step": 385 }, { "epoch": 0.3580705009276438, "grad_norm": 4.739077091217041, "learning_rate": 1.6213169785545688e-05, "loss": 0.2673, "step": 386 }, { "epoch": 0.3589981447124304, "grad_norm": 5.1900634765625, "learning_rate": 1.6187759674112972e-05, "loss": 0.2156, "step": 387 }, { "epoch": 0.35992578849721707, "grad_norm": 2.844031810760498, "learning_rate": 1.6162284656009276e-05, "loss": 0.0855, "step": 388 }, { "epoch": 0.3608534322820037, "grad_norm": 5.831792831420898, "learning_rate": 1.6136744998455477e-05, "loss": 0.2278, "step": 389 }, { "epoch": 0.36178107606679033, "grad_norm": 3.952970504760742, "learning_rate": 1.6111140969350504e-05, "loss": 0.1348, "step": 390 }, { "epoch": 0.362708719851577, "grad_norm": 3.9323179721832275, "learning_rate": 1.6085472837268504e-05, "loss": 0.1731, "step": 391 }, { "epoch": 0.36363636363636365, "grad_norm": 4.621875286102295, "learning_rate": 1.6059740871456035e-05, "loss": 0.2011, "step": 392 }, { "epoch": 0.3645640074211503, "grad_norm": 6.936546325683594, "learning_rate": 1.603394534182925e-05, "loss": 0.1766, "step": 393 }, { "epoch": 0.3654916512059369, "grad_norm": 4.287359237670898, "learning_rate": 1.6008086518971037e-05, "loss": 0.1939, "step": 394 }, { "epoch": 0.36641929499072357, "grad_norm": 3.0315101146698, "learning_rate": 1.598216467412822e-05, "loss": 0.1091, "step": 395 }, { "epoch": 0.3673469387755102, "grad_norm": 6.599478721618652, "learning_rate": 1.5956180079208684e-05, "loss": 0.232, "step": 396 }, { "epoch": 0.3682745825602968, "grad_norm": 4.763383388519287, "learning_rate": 1.593013300677853e-05, "loss": 0.1877, "step": 397 }, { "epoch": 0.3692022263450835, "grad_norm": 4.326437950134277, "learning_rate": 1.5904023730059227e-05, "loss": 0.1924, "step": 398 }, { "epoch": 0.37012987012987014, "grad_norm": 4.332076072692871, "learning_rate": 1.5877852522924733e-05, "loss": 0.201, "step": 399 }, { "epoch": 0.37105751391465674, "grad_norm": 4.557116985321045, "learning_rate": 1.5851619659898623e-05, "loss": 0.2, "step": 400 }, { "epoch": 0.37105751391465674, "eval_accuracy": 0.844789356984479, "eval_f1": 0.6585365853658537, "eval_loss": 0.33066147565841675, "eval_precision": 0.8598726114649682, "eval_recall": 0.5335968379446641, "eval_runtime": 47.9063, "eval_samples_per_second": 5.761, "eval_steps_per_second": 0.188, "step": 400 }, { "epoch": 0.3719851576994434, "grad_norm": 4.4434332847595215, "learning_rate": 1.582532541615122e-05, "loss": 0.1563, "step": 401 }, { "epoch": 0.37291280148423006, "grad_norm": 2.928924083709717, "learning_rate": 1.57989700674967e-05, "loss": 0.1211, "step": 402 }, { "epoch": 0.3738404452690167, "grad_norm": 3.329848527908325, "learning_rate": 1.5772553890390196e-05, "loss": 0.2248, "step": 403 }, { "epoch": 0.3747680890538033, "grad_norm": 4.852266311645508, "learning_rate": 1.5746077161924905e-05, "loss": 0.2121, "step": 404 }, { "epoch": 0.37569573283859, "grad_norm": 8.039565086364746, "learning_rate": 1.5719540159829185e-05, "loss": 0.1864, "step": 405 }, { "epoch": 0.37662337662337664, "grad_norm": 4.178286552429199, "learning_rate": 1.5692943162463628e-05, "loss": 0.1164, "step": 406 }, { "epoch": 0.37755102040816324, "grad_norm": 4.181092739105225, "learning_rate": 1.5666286448818152e-05, "loss": 0.136, "step": 407 }, { "epoch": 0.3784786641929499, "grad_norm": 4.847424030303955, "learning_rate": 1.5639570298509067e-05, "loss": 0.1692, "step": 408 }, { "epoch": 0.37940630797773656, "grad_norm": 5.589816570281982, "learning_rate": 1.5612794991776147e-05, "loss": 0.1375, "step": 409 }, { "epoch": 0.3803339517625232, "grad_norm": 3.174445629119873, "learning_rate": 1.5585960809479698e-05, "loss": 0.1683, "step": 410 }, { "epoch": 0.3812615955473098, "grad_norm": 7.2739739418029785, "learning_rate": 1.5559068033097583e-05, "loss": 0.2135, "step": 411 }, { "epoch": 0.3821892393320965, "grad_norm": 5.514982223510742, "learning_rate": 1.5532116944722308e-05, "loss": 0.2366, "step": 412 }, { "epoch": 0.38311688311688313, "grad_norm": 4.4258036613464355, "learning_rate": 1.5505107827058038e-05, "loss": 0.1667, "step": 413 }, { "epoch": 0.38404452690166974, "grad_norm": 4.089566230773926, "learning_rate": 1.547804096341763e-05, "loss": 0.1647, "step": 414 }, { "epoch": 0.3849721706864564, "grad_norm": 3.6559548377990723, "learning_rate": 1.5450916637719683e-05, "loss": 0.1705, "step": 415 }, { "epoch": 0.38589981447124305, "grad_norm": 4.045246124267578, "learning_rate": 1.5423735134485537e-05, "loss": 0.1915, "step": 416 }, { "epoch": 0.3868274582560297, "grad_norm": 3.6718714237213135, "learning_rate": 1.5396496738836292e-05, "loss": 0.1532, "step": 417 }, { "epoch": 0.3877551020408163, "grad_norm": 4.593350887298584, "learning_rate": 1.536920173648984e-05, "loss": 0.1815, "step": 418 }, { "epoch": 0.38868274582560297, "grad_norm": 3.913667917251587, "learning_rate": 1.5341850413757834e-05, "loss": 0.1448, "step": 419 }, { "epoch": 0.38961038961038963, "grad_norm": 4.148723125457764, "learning_rate": 1.5314443057542703e-05, "loss": 0.2082, "step": 420 }, { "epoch": 0.38961038961038963, "eval_accuracy": 0.8603104212860311, "eval_f1": 0.7136363636363636, "eval_loss": 0.31429344415664673, "eval_precision": 0.839572192513369, "eval_recall": 0.6205533596837944, "eval_runtime": 48.5179, "eval_samples_per_second": 5.689, "eval_steps_per_second": 0.185, "step": 420 }, { "epoch": 0.39053803339517623, "grad_norm": 4.14030122756958, "learning_rate": 1.5286979955334655e-05, "loss": 0.1749, "step": 421 }, { "epoch": 0.3914656771799629, "grad_norm": 3.4345805644989014, "learning_rate": 1.5259461395208628e-05, "loss": 0.1712, "step": 422 }, { "epoch": 0.39239332096474955, "grad_norm": 3.9379382133483887, "learning_rate": 1.52318876658213e-05, "loss": 0.1658, "step": 423 }, { "epoch": 0.39332096474953615, "grad_norm": 3.1555135250091553, "learning_rate": 1.5204259056408046e-05, "loss": 0.1537, "step": 424 }, { "epoch": 0.3942486085343228, "grad_norm": 3.893655776977539, "learning_rate": 1.5176575856779904e-05, "loss": 0.1509, "step": 425 }, { "epoch": 0.39517625231910947, "grad_norm": 4.423066139221191, "learning_rate": 1.5148838357320537e-05, "loss": 0.1501, "step": 426 }, { "epoch": 0.3961038961038961, "grad_norm": 5.078852653503418, "learning_rate": 1.512104684898319e-05, "loss": 0.2835, "step": 427 }, { "epoch": 0.3970315398886827, "grad_norm": 3.377333164215088, "learning_rate": 1.5093201623287631e-05, "loss": 0.1393, "step": 428 }, { "epoch": 0.3979591836734694, "grad_norm": 4.0545196533203125, "learning_rate": 1.5065302972317108e-05, "loss": 0.1567, "step": 429 }, { "epoch": 0.39888682745825604, "grad_norm": 6.253279685974121, "learning_rate": 1.5037351188715265e-05, "loss": 0.2155, "step": 430 }, { "epoch": 0.39981447124304265, "grad_norm": 2.698172092437744, "learning_rate": 1.5009346565683088e-05, "loss": 0.1102, "step": 431 }, { "epoch": 0.4007421150278293, "grad_norm": 4.845376014709473, "learning_rate": 1.4981289396975818e-05, "loss": 0.2186, "step": 432 }, { "epoch": 0.40166975881261596, "grad_norm": 2.9927968978881836, "learning_rate": 1.4953179976899878e-05, "loss": 0.1365, "step": 433 }, { "epoch": 0.4025974025974026, "grad_norm": 3.484133720397949, "learning_rate": 1.4925018600309784e-05, "loss": 0.18, "step": 434 }, { "epoch": 0.4035250463821892, "grad_norm": 2.8350846767425537, "learning_rate": 1.4896805562605052e-05, "loss": 0.1644, "step": 435 }, { "epoch": 0.4044526901669759, "grad_norm": 3.7044577598571777, "learning_rate": 1.4868541159727097e-05, "loss": 0.1758, "step": 436 }, { "epoch": 0.40538033395176254, "grad_norm": 3.7919528484344482, "learning_rate": 1.4840225688156132e-05, "loss": 0.1609, "step": 437 }, { "epoch": 0.40630797773654914, "grad_norm": 3.7263407707214355, "learning_rate": 1.4811859444908053e-05, "loss": 0.1674, "step": 438 }, { "epoch": 0.4072356215213358, "grad_norm": 3.5427684783935547, "learning_rate": 1.4783442727531328e-05, "loss": 0.212, "step": 439 }, { "epoch": 0.40816326530612246, "grad_norm": 5.29186487197876, "learning_rate": 1.4754975834103877e-05, "loss": 0.2051, "step": 440 }, { "epoch": 0.40816326530612246, "eval_accuracy": 0.8569844789356984, "eval_f1": 0.6978922716627635, "eval_loss": 0.3138832747936249, "eval_precision": 0.8563218390804598, "eval_recall": 0.5889328063241107, "eval_runtime": 48.1239, "eval_samples_per_second": 5.735, "eval_steps_per_second": 0.187, "step": 440 }, { "epoch": 0.4090909090909091, "grad_norm": 6.476135730743408, "learning_rate": 1.4726459063229946e-05, "loss": 0.1586, "step": 441 }, { "epoch": 0.4100185528756957, "grad_norm": 4.70169734954834, "learning_rate": 1.4697892714036959e-05, "loss": 0.1604, "step": 442 }, { "epoch": 0.4109461966604824, "grad_norm": 3.1262881755828857, "learning_rate": 1.4669277086172406e-05, "loss": 0.1282, "step": 443 }, { "epoch": 0.41187384044526903, "grad_norm": 3.5137503147125244, "learning_rate": 1.4640612479800686e-05, "loss": 0.179, "step": 444 }, { "epoch": 0.41280148423005564, "grad_norm": 5.388190269470215, "learning_rate": 1.4611899195599952e-05, "loss": 0.2404, "step": 445 }, { "epoch": 0.4137291280148423, "grad_norm": 4.668002128601074, "learning_rate": 1.4583137534758968e-05, "loss": 0.1781, "step": 446 }, { "epoch": 0.41465677179962895, "grad_norm": 4.363613128662109, "learning_rate": 1.455432779897395e-05, "loss": 0.1636, "step": 447 }, { "epoch": 0.4155844155844156, "grad_norm": 3.608628749847412, "learning_rate": 1.4525470290445392e-05, "loss": 0.1604, "step": 448 }, { "epoch": 0.4165120593692022, "grad_norm": 4.1195387840271, "learning_rate": 1.4496565311874902e-05, "loss": 0.1752, "step": 449 }, { "epoch": 0.4174397031539889, "grad_norm": 2.9940760135650635, "learning_rate": 1.4467613166462024e-05, "loss": 0.1579, "step": 450 }, { "epoch": 0.41836734693877553, "grad_norm": 3.3826465606689453, "learning_rate": 1.4438614157901073e-05, "loss": 0.1529, "step": 451 }, { "epoch": 0.41929499072356213, "grad_norm": 2.9892001152038574, "learning_rate": 1.4409568590377918e-05, "loss": 0.1435, "step": 452 }, { "epoch": 0.4202226345083488, "grad_norm": 3.4909090995788574, "learning_rate": 1.4380476768566825e-05, "loss": 0.1879, "step": 453 }, { "epoch": 0.42115027829313545, "grad_norm": 3.8178699016571045, "learning_rate": 1.4351338997627233e-05, "loss": 0.1693, "step": 454 }, { "epoch": 0.42207792207792205, "grad_norm": 3.101930618286133, "learning_rate": 1.4322155583200577e-05, "loss": 0.1315, "step": 455 }, { "epoch": 0.4230055658627087, "grad_norm": 3.344278335571289, "learning_rate": 1.429292683140706e-05, "loss": 0.1539, "step": 456 }, { "epoch": 0.42393320964749537, "grad_norm": 2.970942497253418, "learning_rate": 1.4263653048842461e-05, "loss": 0.1374, "step": 457 }, { "epoch": 0.424860853432282, "grad_norm": 3.1692464351654053, "learning_rate": 1.4234334542574906e-05, "loss": 0.2035, "step": 458 }, { "epoch": 0.42578849721706863, "grad_norm": 2.829171895980835, "learning_rate": 1.4204971620141648e-05, "loss": 0.1483, "step": 459 }, { "epoch": 0.4267161410018553, "grad_norm": 3.1080052852630615, "learning_rate": 1.4175564589545853e-05, "loss": 0.0959, "step": 460 }, { "epoch": 0.4267161410018553, "eval_accuracy": 0.8569844789356984, "eval_f1": 0.6993006993006993, "eval_loss": 0.3129526674747467, "eval_precision": 0.8522727272727273, "eval_recall": 0.5928853754940712, "eval_runtime": 48.0013, "eval_samples_per_second": 5.75, "eval_steps_per_second": 0.187, "step": 460 }, { "epoch": 0.42764378478664195, "grad_norm": 4.785598278045654, "learning_rate": 1.4146113759253362e-05, "loss": 0.1891, "step": 461 }, { "epoch": 0.42857142857142855, "grad_norm": 4.229030609130859, "learning_rate": 1.411661943818944e-05, "loss": 0.1546, "step": 462 }, { "epoch": 0.4294990723562152, "grad_norm": 4.43852424621582, "learning_rate": 1.4087081935735565e-05, "loss": 0.1769, "step": 463 }, { "epoch": 0.43042671614100186, "grad_norm": 4.6244049072265625, "learning_rate": 1.4057501561726157e-05, "loss": 0.2404, "step": 464 }, { "epoch": 0.4313543599257885, "grad_norm": 4.3941168785095215, "learning_rate": 1.4027878626445339e-05, "loss": 0.1781, "step": 465 }, { "epoch": 0.4322820037105751, "grad_norm": 6.995157718658447, "learning_rate": 1.3998213440623691e-05, "loss": 0.2318, "step": 466 }, { "epoch": 0.4332096474953618, "grad_norm": 2.781472682952881, "learning_rate": 1.3968506315434973e-05, "loss": 0.0949, "step": 467 }, { "epoch": 0.43413729128014844, "grad_norm": 2.9485185146331787, "learning_rate": 1.3938757562492873e-05, "loss": 0.1268, "step": 468 }, { "epoch": 0.43506493506493504, "grad_norm": 6.8653130531311035, "learning_rate": 1.390896749384773e-05, "loss": 0.2818, "step": 469 }, { "epoch": 0.4359925788497217, "grad_norm": 5.269331932067871, "learning_rate": 1.3879136421983265e-05, "loss": 0.1386, "step": 470 }, { "epoch": 0.43692022263450836, "grad_norm": 5.839674472808838, "learning_rate": 1.3849264659813314e-05, "loss": 0.1739, "step": 471 }, { "epoch": 0.437847866419295, "grad_norm": 4.306594371795654, "learning_rate": 1.3819352520678519e-05, "loss": 0.1562, "step": 472 }, { "epoch": 0.4387755102040816, "grad_norm": 4.80615234375, "learning_rate": 1.378940031834307e-05, "loss": 0.2049, "step": 473 }, { "epoch": 0.4397031539888683, "grad_norm": 4.824836730957031, "learning_rate": 1.3759408366991391e-05, "loss": 0.1667, "step": 474 }, { "epoch": 0.44063079777365494, "grad_norm": 4.227314472198486, "learning_rate": 1.3729376981224869e-05, "loss": 0.1783, "step": 475 }, { "epoch": 0.44155844155844154, "grad_norm": 4.170743942260742, "learning_rate": 1.3699306476058523e-05, "loss": 0.1724, "step": 476 }, { "epoch": 0.4424860853432282, "grad_norm": 6.098860263824463, "learning_rate": 1.3669197166917723e-05, "loss": 0.2301, "step": 477 }, { "epoch": 0.44341372912801486, "grad_norm": 3.2894742488861084, "learning_rate": 1.3639049369634878e-05, "loss": 0.1636, "step": 478 }, { "epoch": 0.44434137291280146, "grad_norm": 3.913362503051758, "learning_rate": 1.3608863400446113e-05, "loss": 0.1762, "step": 479 }, { "epoch": 0.4452690166975881, "grad_norm": 4.007498741149902, "learning_rate": 1.357863957598796e-05, "loss": 0.1955, "step": 480 }, { "epoch": 0.4452690166975881, "eval_accuracy": 0.8592017738359202, "eval_f1": 0.7080459770114943, "eval_loss": 0.3043546974658966, "eval_precision": 0.8461538461538461, "eval_recall": 0.6086956521739131, "eval_runtime": 48.2191, "eval_samples_per_second": 5.724, "eval_steps_per_second": 0.187, "step": 480 }, { "epoch": 0.4461966604823748, "grad_norm": 4.113521575927734, "learning_rate": 1.3548378213294042e-05, "loss": 0.1875, "step": 481 }, { "epoch": 0.44712430426716143, "grad_norm": 4.532953262329102, "learning_rate": 1.3518079629791725e-05, "loss": 0.1425, "step": 482 }, { "epoch": 0.44805194805194803, "grad_norm": 2.7280144691467285, "learning_rate": 1.3487744143298822e-05, "loss": 0.0783, "step": 483 }, { "epoch": 0.4489795918367347, "grad_norm": 3.562812566757202, "learning_rate": 1.345737207202023e-05, "loss": 0.1569, "step": 484 }, { "epoch": 0.44990723562152135, "grad_norm": 5.523044109344482, "learning_rate": 1.3426963734544601e-05, "loss": 0.2227, "step": 485 }, { "epoch": 0.45083487940630795, "grad_norm": 4.155531883239746, "learning_rate": 1.3396519449841006e-05, "loss": 0.2297, "step": 486 }, { "epoch": 0.4517625231910946, "grad_norm": 2.8421530723571777, "learning_rate": 1.3366039537255589e-05, "loss": 0.1699, "step": 487 }, { "epoch": 0.45269016697588127, "grad_norm": 3.845489263534546, "learning_rate": 1.3335524316508208e-05, "loss": 0.1773, "step": 488 }, { "epoch": 0.4536178107606679, "grad_norm": 3.0157673358917236, "learning_rate": 1.3304974107689088e-05, "loss": 0.1379, "step": 489 }, { "epoch": 0.45454545454545453, "grad_norm": 2.74364972114563, "learning_rate": 1.3274389231255466e-05, "loss": 0.1265, "step": 490 }, { "epoch": 0.4554730983302412, "grad_norm": 6.820478439331055, "learning_rate": 1.3243770008028225e-05, "loss": 0.1849, "step": 491 }, { "epoch": 0.45640074211502785, "grad_norm": 2.741809368133545, "learning_rate": 1.3213116759188525e-05, "loss": 0.1295, "step": 492 }, { "epoch": 0.45732838589981445, "grad_norm": 3.473961353302002, "learning_rate": 1.3182429806274442e-05, "loss": 0.099, "step": 493 }, { "epoch": 0.4582560296846011, "grad_norm": 5.271576881408691, "learning_rate": 1.3151709471177589e-05, "loss": 0.1753, "step": 494 }, { "epoch": 0.45918367346938777, "grad_norm": 6.4196577072143555, "learning_rate": 1.3120956076139746e-05, "loss": 0.2568, "step": 495 }, { "epoch": 0.4601113172541744, "grad_norm": 5.339606761932373, "learning_rate": 1.3090169943749475e-05, "loss": 0.1535, "step": 496 }, { "epoch": 0.461038961038961, "grad_norm": 4.3615593910217285, "learning_rate": 1.305935139693874e-05, "loss": 0.1362, "step": 497 }, { "epoch": 0.4619666048237477, "grad_norm": 4.690557956695557, "learning_rate": 1.3028500758979507e-05, "loss": 0.1776, "step": 498 }, { "epoch": 0.46289424860853434, "grad_norm": 4.32938814163208, "learning_rate": 1.299761835348038e-05, "loss": 0.1429, "step": 499 }, { "epoch": 0.46382189239332094, "grad_norm": 3.8397958278656006, "learning_rate": 1.296670450438317e-05, "loss": 0.1904, "step": 500 }, { "epoch": 0.46382189239332094, "eval_accuracy": 0.8403547671840355, "eval_f1": 0.6381909547738693, "eval_loss": 0.3388740122318268, "eval_precision": 0.8758620689655172, "eval_recall": 0.5019762845849802, "eval_runtime": 48.3395, "eval_samples_per_second": 5.71, "eval_steps_per_second": 0.186, "step": 500 }, { "epoch": 0.4647495361781076, "grad_norm": 4.0140509605407715, "learning_rate": 1.2935759535959528e-05, "loss": 0.2093, "step": 501 }, { "epoch": 0.46567717996289426, "grad_norm": 2.7800872325897217, "learning_rate": 1.2904783772807534e-05, "loss": 0.1333, "step": 502 }, { "epoch": 0.46660482374768086, "grad_norm": 4.866140842437744, "learning_rate": 1.2873777539848284e-05, "loss": 0.2078, "step": 503 }, { "epoch": 0.4675324675324675, "grad_norm": 4.730177879333496, "learning_rate": 1.2842741162322487e-05, "loss": 0.1834, "step": 504 }, { "epoch": 0.4684601113172542, "grad_norm": 3.826939582824707, "learning_rate": 1.2811674965787058e-05, "loss": 0.1712, "step": 505 }, { "epoch": 0.46938775510204084, "grad_norm": 16.19306755065918, "learning_rate": 1.2780579276111702e-05, "loss": 0.2195, "step": 506 }, { "epoch": 0.47031539888682744, "grad_norm": 4.020465850830078, "learning_rate": 1.2749454419475486e-05, "loss": 0.1387, "step": 507 }, { "epoch": 0.4712430426716141, "grad_norm": 3.083519697189331, "learning_rate": 1.2718300722363431e-05, "loss": 0.1595, "step": 508 }, { "epoch": 0.47217068645640076, "grad_norm": 3.982968807220459, "learning_rate": 1.2687118511563075e-05, "loss": 0.1304, "step": 509 }, { "epoch": 0.47309833024118736, "grad_norm": 4.213962554931641, "learning_rate": 1.2655908114161053e-05, "loss": 0.1269, "step": 510 }, { "epoch": 0.474025974025974, "grad_norm": 3.9727156162261963, "learning_rate": 1.2624669857539669e-05, "loss": 0.1327, "step": 511 }, { "epoch": 0.4749536178107607, "grad_norm": 4.792732238769531, "learning_rate": 1.2593404069373452e-05, "loss": 0.1782, "step": 512 }, { "epoch": 0.47588126159554733, "grad_norm": 3.0668811798095703, "learning_rate": 1.2562111077625723e-05, "loss": 0.1568, "step": 513 }, { "epoch": 0.47680890538033394, "grad_norm": 4.204139709472656, "learning_rate": 1.2530791210545163e-05, "loss": 0.149, "step": 514 }, { "epoch": 0.4777365491651206, "grad_norm": 4.396578788757324, "learning_rate": 1.2499444796662354e-05, "loss": 0.1943, "step": 515 }, { "epoch": 0.47866419294990725, "grad_norm": 8.477376937866211, "learning_rate": 1.2468072164786342e-05, "loss": 0.3153, "step": 516 }, { "epoch": 0.47959183673469385, "grad_norm": 4.493762016296387, "learning_rate": 1.2436673644001196e-05, "loss": 0.2028, "step": 517 }, { "epoch": 0.4805194805194805, "grad_norm": 3.7687320709228516, "learning_rate": 1.2405249563662539e-05, "loss": 0.1834, "step": 518 }, { "epoch": 0.48144712430426717, "grad_norm": 5.828054428100586, "learning_rate": 1.23738002533941e-05, "loss": 0.1587, "step": 519 }, { "epoch": 0.48237476808905383, "grad_norm": 4.38277006149292, "learning_rate": 1.2342326043084268e-05, "loss": 0.1809, "step": 520 }, { "epoch": 0.48237476808905383, "eval_accuracy": 0.8458980044345898, "eval_f1": 0.6584766584766585, "eval_loss": 0.3318649232387543, "eval_precision": 0.8701298701298701, "eval_recall": 0.5296442687747036, "eval_runtime": 46.7049, "eval_samples_per_second": 5.909, "eval_steps_per_second": 0.193, "step": 520 }, { "epoch": 0.48330241187384043, "grad_norm": 4.514082908630371, "learning_rate": 1.2310827262882614e-05, "loss": 0.1759, "step": 521 }, { "epoch": 0.4842300556586271, "grad_norm": 3.4471170902252197, "learning_rate": 1.2279304243196438e-05, "loss": 0.1364, "step": 522 }, { "epoch": 0.48515769944341375, "grad_norm": 2.6771576404571533, "learning_rate": 1.2247757314687296e-05, "loss": 0.146, "step": 523 }, { "epoch": 0.48608534322820035, "grad_norm": 3.6538710594177246, "learning_rate": 1.2216186808267544e-05, "loss": 0.2009, "step": 524 }, { "epoch": 0.487012987012987, "grad_norm": 4.4636149406433105, "learning_rate": 1.2184593055096853e-05, "loss": 0.1474, "step": 525 }, { "epoch": 0.48794063079777367, "grad_norm": 5.249742031097412, "learning_rate": 1.215297638657875e-05, "loss": 0.1564, "step": 526 }, { "epoch": 0.48886827458256027, "grad_norm": 6.001680850982666, "learning_rate": 1.2121337134357121e-05, "loss": 0.1718, "step": 527 }, { "epoch": 0.4897959183673469, "grad_norm": 6.546911716461182, "learning_rate": 1.2089675630312755e-05, "loss": 0.2193, "step": 528 }, { "epoch": 0.4907235621521336, "grad_norm": 2.6112513542175293, "learning_rate": 1.2057992206559837e-05, "loss": 0.1295, "step": 529 }, { "epoch": 0.49165120593692024, "grad_norm": 3.9656717777252197, "learning_rate": 1.2026287195442503e-05, "loss": 0.1707, "step": 530 }, { "epoch": 0.49257884972170685, "grad_norm": 3.8426477909088135, "learning_rate": 1.199456092953131e-05, "loss": 0.1768, "step": 531 }, { "epoch": 0.4935064935064935, "grad_norm": 3.001831531524658, "learning_rate": 1.1962813741619777e-05, "loss": 0.1839, "step": 532 }, { "epoch": 0.49443413729128016, "grad_norm": 5.149347305297852, "learning_rate": 1.1931045964720882e-05, "loss": 0.2559, "step": 533 }, { "epoch": 0.49536178107606677, "grad_norm": 3.0468552112579346, "learning_rate": 1.189925793206357e-05, "loss": 0.1408, "step": 534 }, { "epoch": 0.4962894248608534, "grad_norm": 4.5860490798950195, "learning_rate": 1.1867449977089264e-05, "loss": 0.1945, "step": 535 }, { "epoch": 0.4972170686456401, "grad_norm": 3.8901429176330566, "learning_rate": 1.1835622433448361e-05, "loss": 0.2126, "step": 536 }, { "epoch": 0.49814471243042674, "grad_norm": 4.597271919250488, "learning_rate": 1.1803775634996735e-05, "loss": 0.1977, "step": 537 }, { "epoch": 0.49907235621521334, "grad_norm": 3.079770803451538, "learning_rate": 1.177190991579223e-05, "loss": 0.1758, "step": 538 }, { "epoch": 0.5, "grad_norm": 4.214216709136963, "learning_rate": 1.174002561009116e-05, "loss": 0.141, "step": 539 }, { "epoch": 0.5009276437847866, "grad_norm": 5.213557243347168, "learning_rate": 1.1708123052344803e-05, "loss": 0.1605, "step": 540 }, { "epoch": 0.5009276437847866, "eval_accuracy": 0.8614190687361419, "eval_f1": 0.7072599531615925, "eval_loss": 0.30164089798927307, "eval_precision": 0.867816091954023, "eval_recall": 0.5968379446640316, "eval_runtime": 47.208, "eval_samples_per_second": 5.846, "eval_steps_per_second": 0.191, "step": 540 }, { "epoch": 0.5018552875695733, "grad_norm": 2.1639671325683594, "learning_rate": 1.1676202577195901e-05, "loss": 0.0809, "step": 541 }, { "epoch": 0.5027829313543599, "grad_norm": 4.442990303039551, "learning_rate": 1.164426451947513e-05, "loss": 0.1677, "step": 542 }, { "epoch": 0.5037105751391465, "grad_norm": 3.968435049057007, "learning_rate": 1.1612309214197599e-05, "loss": 0.2114, "step": 543 }, { "epoch": 0.5046382189239332, "grad_norm": 9.67383098602295, "learning_rate": 1.1580336996559343e-05, "loss": 0.2159, "step": 544 }, { "epoch": 0.5055658627087198, "grad_norm": 3.427710771560669, "learning_rate": 1.1548348201933799e-05, "loss": 0.0846, "step": 545 }, { "epoch": 0.5064935064935064, "grad_norm": 3.03241229057312, "learning_rate": 1.151634316586828e-05, "loss": 0.1195, "step": 546 }, { "epoch": 0.5074211502782932, "grad_norm": 5.138626575469971, "learning_rate": 1.1484322224080474e-05, "loss": 0.1542, "step": 547 }, { "epoch": 0.5083487940630798, "grad_norm": 4.37513542175293, "learning_rate": 1.1452285712454905e-05, "loss": 0.2118, "step": 548 }, { "epoch": 0.5092764378478665, "grad_norm": 4.746356010437012, "learning_rate": 1.1420233967039423e-05, "loss": 0.1456, "step": 549 }, { "epoch": 0.5102040816326531, "grad_norm": 5.190648555755615, "learning_rate": 1.138816732404167e-05, "loss": 0.1921, "step": 550 }, { "epoch": 0.5111317254174397, "grad_norm": 3.955061197280884, "learning_rate": 1.1356086119825553e-05, "loss": 0.1964, "step": 551 }, { "epoch": 0.5120593692022264, "grad_norm": 7.773352146148682, "learning_rate": 1.1323990690907734e-05, "loss": 0.2178, "step": 552 }, { "epoch": 0.512987012987013, "grad_norm": 3.211651563644409, "learning_rate": 1.1291881373954066e-05, "loss": 0.1859, "step": 553 }, { "epoch": 0.5139146567717996, "grad_norm": 3.536742687225342, "learning_rate": 1.1259758505776092e-05, "loss": 0.1949, "step": 554 }, { "epoch": 0.5148423005565863, "grad_norm": 4.817080974578857, "learning_rate": 1.1227622423327501e-05, "loss": 0.2482, "step": 555 }, { "epoch": 0.5157699443413729, "grad_norm": 4.828971862792969, "learning_rate": 1.119547346370059e-05, "loss": 0.216, "step": 556 }, { "epoch": 0.5166975881261595, "grad_norm": 4.580413818359375, "learning_rate": 1.1163311964122733e-05, "loss": 0.2267, "step": 557 }, { "epoch": 0.5176252319109462, "grad_norm": 2.9997427463531494, "learning_rate": 1.1131138261952845e-05, "loss": 0.1556, "step": 558 }, { "epoch": 0.5185528756957328, "grad_norm": 2.876847982406616, "learning_rate": 1.109895269467783e-05, "loss": 0.1551, "step": 559 }, { "epoch": 0.5194805194805194, "grad_norm": 6.003294467926025, "learning_rate": 1.1066755599909065e-05, "loss": 0.2123, "step": 560 }, { "epoch": 0.5194805194805194, "eval_accuracy": 0.8603104212860311, "eval_f1": 0.7136363636363636, "eval_loss": 0.2982672452926636, "eval_precision": 0.839572192513369, "eval_recall": 0.6205533596837944, "eval_runtime": 48.6549, "eval_samples_per_second": 5.673, "eval_steps_per_second": 0.185, "step": 560 }, { "epoch": 0.5204081632653061, "grad_norm": 3.7513365745544434, "learning_rate": 1.1034547315378838e-05, "loss": 0.1808, "step": 561 }, { "epoch": 0.5213358070500927, "grad_norm": 2.755664110183716, "learning_rate": 1.1002328178936813e-05, "loss": 0.1272, "step": 562 }, { "epoch": 0.5222634508348795, "grad_norm": 6.458003520965576, "learning_rate": 1.0970098528546482e-05, "loss": 0.161, "step": 563 }, { "epoch": 0.5231910946196661, "grad_norm": 3.156869649887085, "learning_rate": 1.0937858702281631e-05, "loss": 0.1423, "step": 564 }, { "epoch": 0.5241187384044527, "grad_norm": 4.826041221618652, "learning_rate": 1.090560903832278e-05, "loss": 0.1717, "step": 565 }, { "epoch": 0.5250463821892394, "grad_norm": 2.8493692874908447, "learning_rate": 1.087334987495364e-05, "loss": 0.147, "step": 566 }, { "epoch": 0.525974025974026, "grad_norm": 3.3068854808807373, "learning_rate": 1.0841081550557577e-05, "loss": 0.1084, "step": 567 }, { "epoch": 0.5269016697588126, "grad_norm": 3.908871650695801, "learning_rate": 1.0808804403614044e-05, "loss": 0.1484, "step": 568 }, { "epoch": 0.5278293135435993, "grad_norm": 3.8984129428863525, "learning_rate": 1.0776518772695035e-05, "loss": 0.1883, "step": 569 }, { "epoch": 0.5287569573283859, "grad_norm": 3.58467960357666, "learning_rate": 1.0744224996461541e-05, "loss": 0.1238, "step": 570 }, { "epoch": 0.5296846011131725, "grad_norm": 5.206255912780762, "learning_rate": 1.0711923413659995e-05, "loss": 0.2008, "step": 571 }, { "epoch": 0.5306122448979592, "grad_norm": 2.4864425659179688, "learning_rate": 1.0679614363118718e-05, "loss": 0.1181, "step": 572 }, { "epoch": 0.5315398886827458, "grad_norm": 3.949312686920166, "learning_rate": 1.0647298183744359e-05, "loss": 0.1927, "step": 573 }, { "epoch": 0.5324675324675324, "grad_norm": 6.005074501037598, "learning_rate": 1.061497521451835e-05, "loss": 0.1936, "step": 574 }, { "epoch": 0.5333951762523191, "grad_norm": 4.429588317871094, "learning_rate": 1.0582645794493337e-05, "loss": 0.2031, "step": 575 }, { "epoch": 0.5343228200371057, "grad_norm": 4.36995792388916, "learning_rate": 1.055031026278965e-05, "loss": 0.1846, "step": 576 }, { "epoch": 0.5352504638218923, "grad_norm": 4.059164524078369, "learning_rate": 1.0517968958591705e-05, "loss": 0.1681, "step": 577 }, { "epoch": 0.536178107606679, "grad_norm": 4.211386203765869, "learning_rate": 1.0485622221144485e-05, "loss": 0.1296, "step": 578 }, { "epoch": 0.5371057513914657, "grad_norm": 4.114365100860596, "learning_rate": 1.0453270389749956e-05, "loss": 0.164, "step": 579 }, { "epoch": 0.5380333951762524, "grad_norm": 7.943957805633545, "learning_rate": 1.0420913803763522e-05, "loss": 0.2279, "step": 580 }, { "epoch": 0.5380333951762524, "eval_accuracy": 0.8558758314855875, "eval_f1": 0.7018348623853211, "eval_loss": 0.30458346009254456, "eval_precision": 0.8360655737704918, "eval_recall": 0.6047430830039525, "eval_runtime": 46.8877, "eval_samples_per_second": 5.886, "eval_steps_per_second": 0.192, "step": 580 }, { "epoch": 0.538961038961039, "grad_norm": 7.277109146118164, "learning_rate": 1.0388552802590461e-05, "loss": 0.0867, "step": 581 }, { "epoch": 0.5398886827458256, "grad_norm": 3.722276449203491, "learning_rate": 1.0356187725682359e-05, "loss": 0.1233, "step": 582 }, { "epoch": 0.5408163265306123, "grad_norm": 4.75911283493042, "learning_rate": 1.0323818912533561e-05, "loss": 0.2018, "step": 583 }, { "epoch": 0.5417439703153989, "grad_norm": 5.309391498565674, "learning_rate": 1.0291446702677598e-05, "loss": 0.1853, "step": 584 }, { "epoch": 0.5426716141001855, "grad_norm": 3.6361756324768066, "learning_rate": 1.0259071435683636e-05, "loss": 0.1594, "step": 585 }, { "epoch": 0.5435992578849722, "grad_norm": 8.379057884216309, "learning_rate": 1.02266934511529e-05, "loss": 0.1496, "step": 586 }, { "epoch": 0.5445269016697588, "grad_norm": 6.6115593910217285, "learning_rate": 1.0194313088715135e-05, "loss": 0.1887, "step": 587 }, { "epoch": 0.5454545454545454, "grad_norm": 3.5273900032043457, "learning_rate": 1.0161930688025018e-05, "loss": 0.1549, "step": 588 }, { "epoch": 0.5463821892393321, "grad_norm": 4.194179058074951, "learning_rate": 1.0129546588758605e-05, "loss": 0.1748, "step": 589 }, { "epoch": 0.5473098330241187, "grad_norm": 3.131457805633545, "learning_rate": 1.0097161130609774e-05, "loss": 0.1319, "step": 590 }, { "epoch": 0.5482374768089053, "grad_norm": 4.083921909332275, "learning_rate": 1.0064774653286662e-05, "loss": 0.1759, "step": 591 }, { "epoch": 0.549165120593692, "grad_norm": 3.384917736053467, "learning_rate": 1.003238749650809e-05, "loss": 0.1434, "step": 592 }, { "epoch": 0.5500927643784786, "grad_norm": 2.5680859088897705, "learning_rate": 1e-05, "loss": 0.1626, "step": 593 }, { "epoch": 0.5510204081632653, "grad_norm": 3.510484457015991, "learning_rate": 9.967612503491915e-06, "loss": 0.1701, "step": 594 }, { "epoch": 0.551948051948052, "grad_norm": 5.625200271606445, "learning_rate": 9.935225346713341e-06, "loss": 0.1486, "step": 595 }, { "epoch": 0.5528756957328386, "grad_norm": 5.352198123931885, "learning_rate": 9.90283886939023e-06, "loss": 0.1158, "step": 596 }, { "epoch": 0.5538033395176253, "grad_norm": 3.0487513542175293, "learning_rate": 9.870453411241399e-06, "loss": 0.1339, "step": 597 }, { "epoch": 0.5547309833024119, "grad_norm": 3.2191503047943115, "learning_rate": 9.838069311974986e-06, "loss": 0.1147, "step": 598 }, { "epoch": 0.5556586270871985, "grad_norm": 4.020503520965576, "learning_rate": 9.805686911284867e-06, "loss": 0.1163, "step": 599 }, { "epoch": 0.5565862708719852, "grad_norm": 6.643052101135254, "learning_rate": 9.773306548847102e-06, "loss": 0.2224, "step": 600 }, { "epoch": 0.5565862708719852, "eval_accuracy": 0.8381374722838137, "eval_f1": 0.6313131313131313, "eval_loss": 0.3394555449485779, "eval_precision": 0.8741258741258742, "eval_recall": 0.49407114624505927, "eval_runtime": 48.1127, "eval_samples_per_second": 5.737, "eval_steps_per_second": 0.187, "step": 600 }, { "epoch": 0.5575139146567718, "grad_norm": 4.177931308746338, "learning_rate": 9.740928564316369e-06, "loss": 0.1734, "step": 601 }, { "epoch": 0.5584415584415584, "grad_norm": 6.294393062591553, "learning_rate": 9.708553297322407e-06, "loss": 0.1923, "step": 602 }, { "epoch": 0.5593692022263451, "grad_norm": 3.275909662246704, "learning_rate": 9.676181087466444e-06, "loss": 0.1662, "step": 603 }, { "epoch": 0.5602968460111317, "grad_norm": 3.7851905822753906, "learning_rate": 9.643812274317644e-06, "loss": 0.1521, "step": 604 }, { "epoch": 0.5612244897959183, "grad_norm": 4.724743366241455, "learning_rate": 9.611447197409544e-06, "loss": 0.1713, "step": 605 }, { "epoch": 0.562152133580705, "grad_norm": 6.553765773773193, "learning_rate": 9.579086196236483e-06, "loss": 0.1596, "step": 606 }, { "epoch": 0.5630797773654916, "grad_norm": 3.5295910835266113, "learning_rate": 9.54672961025005e-06, "loss": 0.1436, "step": 607 }, { "epoch": 0.5640074211502782, "grad_norm": 4.828715801239014, "learning_rate": 9.514377778855521e-06, "loss": 0.2361, "step": 608 }, { "epoch": 0.564935064935065, "grad_norm": 6.0731072425842285, "learning_rate": 9.482031041408296e-06, "loss": 0.165, "step": 609 }, { "epoch": 0.5658627087198516, "grad_norm": 8.083799362182617, "learning_rate": 9.449689737210352e-06, "loss": 0.1995, "step": 610 }, { "epoch": 0.5667903525046383, "grad_norm": 3.7561042308807373, "learning_rate": 9.417354205506663e-06, "loss": 0.214, "step": 611 }, { "epoch": 0.5677179962894249, "grad_norm": 4.848057746887207, "learning_rate": 9.385024785481653e-06, "loss": 0.2019, "step": 612 }, { "epoch": 0.5686456400742115, "grad_norm": 5.188536167144775, "learning_rate": 9.352701816255643e-06, "loss": 0.1568, "step": 613 }, { "epoch": 0.5695732838589982, "grad_norm": 2.897571325302124, "learning_rate": 9.320385636881283e-06, "loss": 0.1241, "step": 614 }, { "epoch": 0.5705009276437848, "grad_norm": 5.466006278991699, "learning_rate": 9.288076586340005e-06, "loss": 0.195, "step": 615 }, { "epoch": 0.5714285714285714, "grad_norm": 3.484766960144043, "learning_rate": 9.255775003538462e-06, "loss": 0.1419, "step": 616 }, { "epoch": 0.5723562152133581, "grad_norm": 5.067159175872803, "learning_rate": 9.22348122730497e-06, "loss": 0.1938, "step": 617 }, { "epoch": 0.5732838589981447, "grad_norm": 4.161076545715332, "learning_rate": 9.19119559638596e-06, "loss": 0.1925, "step": 618 }, { "epoch": 0.5742115027829313, "grad_norm": 3.2418394088745117, "learning_rate": 9.158918449442425e-06, "loss": 0.1592, "step": 619 }, { "epoch": 0.575139146567718, "grad_norm": 5.799723148345947, "learning_rate": 9.126650125046361e-06, "loss": 0.1655, "step": 620 }, { "epoch": 0.575139146567718, "eval_accuracy": 0.835920177383592, "eval_f1": 0.6224489795918368, "eval_loss": 0.3387525677680969, "eval_precision": 0.8776978417266187, "eval_recall": 0.48221343873517786, "eval_runtime": 47.1826, "eval_samples_per_second": 5.85, "eval_steps_per_second": 0.191, "step": 620 }, { "epoch": 0.5760667903525046, "grad_norm": 4.049962997436523, "learning_rate": 9.094390961677223e-06, "loss": 0.1364, "step": 621 }, { "epoch": 0.5769944341372912, "grad_norm": 3.7945103645324707, "learning_rate": 9.062141297718372e-06, "loss": 0.18, "step": 622 }, { "epoch": 0.577922077922078, "grad_norm": 3.9919049739837646, "learning_rate": 9.02990147145352e-06, "loss": 0.1833, "step": 623 }, { "epoch": 0.5788497217068646, "grad_norm": 2.300546407699585, "learning_rate": 8.99767182106319e-06, "loss": 0.0625, "step": 624 }, { "epoch": 0.5797773654916512, "grad_norm": 4.982729911804199, "learning_rate": 8.965452684621164e-06, "loss": 0.1884, "step": 625 }, { "epoch": 0.5807050092764379, "grad_norm": 4.047850131988525, "learning_rate": 8.933244400090937e-06, "loss": 0.1206, "step": 626 }, { "epoch": 0.5816326530612245, "grad_norm": 4.123579978942871, "learning_rate": 8.901047305322172e-06, "loss": 0.1503, "step": 627 }, { "epoch": 0.5825602968460112, "grad_norm": 4.2010273933410645, "learning_rate": 8.868861738047158e-06, "loss": 0.1311, "step": 628 }, { "epoch": 0.5834879406307978, "grad_norm": 4.511427402496338, "learning_rate": 8.836688035877268e-06, "loss": 0.1521, "step": 629 }, { "epoch": 0.5844155844155844, "grad_norm": 5.0462775230407715, "learning_rate": 8.804526536299413e-06, "loss": 0.1746, "step": 630 }, { "epoch": 0.5853432282003711, "grad_norm": 6.308788299560547, "learning_rate": 8.772377576672502e-06, "loss": 0.2369, "step": 631 }, { "epoch": 0.5862708719851577, "grad_norm": 4.165151596069336, "learning_rate": 8.740241494223911e-06, "loss": 0.2034, "step": 632 }, { "epoch": 0.5871985157699443, "grad_norm": 4.502403259277344, "learning_rate": 8.708118626045939e-06, "loss": 0.1893, "step": 633 }, { "epoch": 0.588126159554731, "grad_norm": 3.4875142574310303, "learning_rate": 8.676009309092273e-06, "loss": 0.1328, "step": 634 }, { "epoch": 0.5890538033395176, "grad_norm": 4.179616928100586, "learning_rate": 8.643913880174449e-06, "loss": 0.1466, "step": 635 }, { "epoch": 0.5899814471243042, "grad_norm": 3.073174476623535, "learning_rate": 8.611832675958335e-06, "loss": 0.1269, "step": 636 }, { "epoch": 0.5909090909090909, "grad_norm": 3.9491851329803467, "learning_rate": 8.579766032960582e-06, "loss": 0.1247, "step": 637 }, { "epoch": 0.5918367346938775, "grad_norm": 3.502418041229248, "learning_rate": 8.5477142875451e-06, "loss": 0.1536, "step": 638 }, { "epoch": 0.5927643784786641, "grad_norm": 5.294571876525879, "learning_rate": 8.515677775919528e-06, "loss": 0.213, "step": 639 }, { "epoch": 0.5936920222634509, "grad_norm": 3.3282008171081543, "learning_rate": 8.48365683413172e-06, "loss": 0.1468, "step": 640 }, { "epoch": 0.5936920222634509, "eval_accuracy": 0.8592017738359202, "eval_f1": 0.7093821510297483, "eval_loss": 0.3021511137485504, "eval_precision": 0.842391304347826, "eval_recall": 0.6126482213438735, "eval_runtime": 46.8153, "eval_samples_per_second": 5.896, "eval_steps_per_second": 0.192, "step": 640 }, { "epoch": 0.5946196660482375, "grad_norm": 5.158597946166992, "learning_rate": 8.451651798066203e-06, "loss": 0.2459, "step": 641 }, { "epoch": 0.5955473098330241, "grad_norm": 7.210265159606934, "learning_rate": 8.419663003440657e-06, "loss": 0.1745, "step": 642 }, { "epoch": 0.5964749536178108, "grad_norm": 4.469946384429932, "learning_rate": 8.387690785802403e-06, "loss": 0.1879, "step": 643 }, { "epoch": 0.5974025974025974, "grad_norm": 3.5130228996276855, "learning_rate": 8.355735480524874e-06, "loss": 0.1453, "step": 644 }, { "epoch": 0.5983302411873841, "grad_norm": 3.8892977237701416, "learning_rate": 8.3237974228041e-06, "loss": 0.1852, "step": 645 }, { "epoch": 0.5992578849721707, "grad_norm": 4.912310600280762, "learning_rate": 8.291876947655197e-06, "loss": 0.1595, "step": 646 }, { "epoch": 0.6001855287569573, "grad_norm": 3.8313469886779785, "learning_rate": 8.259974389908842e-06, "loss": 0.1207, "step": 647 }, { "epoch": 0.601113172541744, "grad_norm": 4.376918792724609, "learning_rate": 8.228090084207773e-06, "loss": 0.2188, "step": 648 }, { "epoch": 0.6020408163265306, "grad_norm": 4.086894989013672, "learning_rate": 8.196224365003267e-06, "loss": 0.1806, "step": 649 }, { "epoch": 0.6029684601113172, "grad_norm": 2.9185261726379395, "learning_rate": 8.16437756655164e-06, "loss": 0.1301, "step": 650 }, { "epoch": 0.6038961038961039, "grad_norm": 5.638646125793457, "learning_rate": 8.132550022910737e-06, "loss": 0.2069, "step": 651 }, { "epoch": 0.6048237476808905, "grad_norm": 6.044648170471191, "learning_rate": 8.100742067936432e-06, "loss": 0.2543, "step": 652 }, { "epoch": 0.6057513914656771, "grad_norm": 5.261450290679932, "learning_rate": 8.068954035279121e-06, "loss": 0.1861, "step": 653 }, { "epoch": 0.6066790352504638, "grad_norm": 2.6060245037078857, "learning_rate": 8.037186258380226e-06, "loss": 0.1102, "step": 654 }, { "epoch": 0.6076066790352505, "grad_norm": 5.192462921142578, "learning_rate": 8.005439070468692e-06, "loss": 0.1196, "step": 655 }, { "epoch": 0.608534322820037, "grad_norm": 4.267360210418701, "learning_rate": 7.9737128045575e-06, "loss": 0.1549, "step": 656 }, { "epoch": 0.6094619666048238, "grad_norm": 2.9216721057891846, "learning_rate": 7.942007793440165e-06, "loss": 0.1326, "step": 657 }, { "epoch": 0.6103896103896104, "grad_norm": 3.592209577560425, "learning_rate": 7.91032436968725e-06, "loss": 0.1399, "step": 658 }, { "epoch": 0.6113172541743971, "grad_norm": 3.6594440937042236, "learning_rate": 7.87866286564288e-06, "loss": 0.1223, "step": 659 }, { "epoch": 0.6122448979591837, "grad_norm": 6.019983291625977, "learning_rate": 7.847023613421251e-06, "loss": 0.1421, "step": 660 }, { "epoch": 0.6122448979591837, "eval_accuracy": 0.843680709534368, "eval_f1": 0.6483790523690773, "eval_loss": 0.3296959400177002, "eval_precision": 0.8783783783783784, "eval_recall": 0.5138339920948617, "eval_runtime": 47.6918, "eval_samples_per_second": 5.787, "eval_steps_per_second": 0.189, "step": 660 }, { "epoch": 0.6131725417439703, "grad_norm": 3.82082462310791, "learning_rate": 7.815406944903148e-06, "loss": 0.1369, "step": 661 }, { "epoch": 0.614100185528757, "grad_norm": 3.763209819793701, "learning_rate": 7.78381319173246e-06, "loss": 0.1345, "step": 662 }, { "epoch": 0.6150278293135436, "grad_norm": 5.018528938293457, "learning_rate": 7.752242685312709e-06, "loss": 0.1762, "step": 663 }, { "epoch": 0.6159554730983302, "grad_norm": 3.500230073928833, "learning_rate": 7.720695756803569e-06, "loss": 0.0632, "step": 664 }, { "epoch": 0.6168831168831169, "grad_norm": 4.73206090927124, "learning_rate": 7.689172737117389e-06, "loss": 0.1771, "step": 665 }, { "epoch": 0.6178107606679035, "grad_norm": 5.639405250549316, "learning_rate": 7.657673956915735e-06, "loss": 0.2544, "step": 666 }, { "epoch": 0.6187384044526901, "grad_norm": 4.007811546325684, "learning_rate": 7.6261997466059035e-06, "loss": 0.1547, "step": 667 }, { "epoch": 0.6196660482374768, "grad_norm": 4.930268287658691, "learning_rate": 7.594750436337467e-06, "loss": 0.1426, "step": 668 }, { "epoch": 0.6205936920222634, "grad_norm": 5.008491516113281, "learning_rate": 7.5633263559988035e-06, "loss": 0.1129, "step": 669 }, { "epoch": 0.62152133580705, "grad_norm": 3.0749435424804688, "learning_rate": 7.531927835213657e-06, "loss": 0.1166, "step": 670 }, { "epoch": 0.6224489795918368, "grad_norm": 5.739880084991455, "learning_rate": 7.500555203337647e-06, "loss": 0.1634, "step": 671 }, { "epoch": 0.6233766233766234, "grad_norm": 3.3177554607391357, "learning_rate": 7.469208789454838e-06, "loss": 0.1388, "step": 672 }, { "epoch": 0.62430426716141, "grad_norm": 4.005073070526123, "learning_rate": 7.4378889223742766e-06, "loss": 0.1429, "step": 673 }, { "epoch": 0.6252319109461967, "grad_norm": 4.268346786499023, "learning_rate": 7.40659593062655e-06, "loss": 0.187, "step": 674 }, { "epoch": 0.6261595547309833, "grad_norm": 6.763199329376221, "learning_rate": 7.375330142460331e-06, "loss": 0.2241, "step": 675 }, { "epoch": 0.62708719851577, "grad_norm": 3.3440911769866943, "learning_rate": 7.344091885838949e-06, "loss": 0.1407, "step": 676 }, { "epoch": 0.6280148423005566, "grad_norm": 5.204977035522461, "learning_rate": 7.312881488436928e-06, "loss": 0.2322, "step": 677 }, { "epoch": 0.6289424860853432, "grad_norm": 4.433789253234863, "learning_rate": 7.2816992776365714e-06, "loss": 0.1657, "step": 678 }, { "epoch": 0.6298701298701299, "grad_norm": 4.035289287567139, "learning_rate": 7.250545580524515e-06, "loss": 0.1994, "step": 679 }, { "epoch": 0.6307977736549165, "grad_norm": 4.986621856689453, "learning_rate": 7.219420723888301e-06, "loss": 0.2483, "step": 680 }, { "epoch": 0.6307977736549165, "eval_accuracy": 0.852549889135255, "eval_f1": 0.6855791962174941, "eval_loss": 0.30598554015159607, "eval_precision": 0.8529411764705882, "eval_recall": 0.5731225296442688, "eval_runtime": 46.8299, "eval_samples_per_second": 5.894, "eval_steps_per_second": 0.192, "step": 680 }, { "epoch": 0.6317254174397031, "grad_norm": 5.635754108428955, "learning_rate": 7.188325034212944e-06, "loss": 0.2133, "step": 681 }, { "epoch": 0.6326530612244898, "grad_norm": 6.280815601348877, "learning_rate": 7.157258837677514e-06, "loss": 0.2312, "step": 682 }, { "epoch": 0.6335807050092764, "grad_norm": 2.508446455001831, "learning_rate": 7.126222460151719e-06, "loss": 0.1189, "step": 683 }, { "epoch": 0.634508348794063, "grad_norm": 3.8485682010650635, "learning_rate": 7.095216227192467e-06, "loss": 0.1505, "step": 684 }, { "epoch": 0.6354359925788498, "grad_norm": 3.7377712726593018, "learning_rate": 7.064240464040472e-06, "loss": 0.129, "step": 685 }, { "epoch": 0.6363636363636364, "grad_norm": 3.6466944217681885, "learning_rate": 7.033295495616834e-06, "loss": 0.1457, "step": 686 }, { "epoch": 0.637291280148423, "grad_norm": 4.0310564041137695, "learning_rate": 7.002381646519625e-06, "loss": 0.1739, "step": 687 }, { "epoch": 0.6382189239332097, "grad_norm": 6.287572860717773, "learning_rate": 6.971499241020495e-06, "loss": 0.1513, "step": 688 }, { "epoch": 0.6391465677179963, "grad_norm": 5.086080551147461, "learning_rate": 6.940648603061263e-06, "loss": 0.2203, "step": 689 }, { "epoch": 0.640074211502783, "grad_norm": 3.155775785446167, "learning_rate": 6.909830056250527e-06, "loss": 0.1288, "step": 690 }, { "epoch": 0.6410018552875696, "grad_norm": 4.002538204193115, "learning_rate": 6.8790439238602576e-06, "loss": 0.1526, "step": 691 }, { "epoch": 0.6419294990723562, "grad_norm": 5.535161972045898, "learning_rate": 6.848290528822417e-06, "loss": 0.2223, "step": 692 }, { "epoch": 0.6428571428571429, "grad_norm": 4.103289604187012, "learning_rate": 6.8175701937255645e-06, "loss": 0.1831, "step": 693 }, { "epoch": 0.6437847866419295, "grad_norm": 4.057906150817871, "learning_rate": 6.786883240811479e-06, "loss": 0.1619, "step": 694 }, { "epoch": 0.6447124304267161, "grad_norm": 2.6124215126037598, "learning_rate": 6.756229991971779e-06, "loss": 0.0988, "step": 695 }, { "epoch": 0.6456400742115028, "grad_norm": 3.484631061553955, "learning_rate": 6.725610768744535e-06, "loss": 0.1127, "step": 696 }, { "epoch": 0.6465677179962894, "grad_norm": 3.0936501026153564, "learning_rate": 6.695025892310913e-06, "loss": 0.1455, "step": 697 }, { "epoch": 0.647495361781076, "grad_norm": 3.7610599994659424, "learning_rate": 6.664475683491797e-06, "loss": 0.1725, "step": 698 }, { "epoch": 0.6484230055658627, "grad_norm": 3.537524700164795, "learning_rate": 6.633960462744415e-06, "loss": 0.1971, "step": 699 }, { "epoch": 0.6493506493506493, "grad_norm": 4.47899055480957, "learning_rate": 6.603480550158995e-06, "loss": 0.1411, "step": 700 }, { "epoch": 0.6493506493506493, "eval_accuracy": 0.8481152993348116, "eval_f1": 0.6714628297362111, "eval_loss": 0.31714314222335815, "eval_precision": 0.8536585365853658, "eval_recall": 0.5533596837944664, "eval_runtime": 47.2966, "eval_samples_per_second": 5.836, "eval_steps_per_second": 0.19, "step": 700 }, { "epoch": 0.650278293135436, "grad_norm": 5.3951334953308105, "learning_rate": 6.5730362654554015e-06, "loss": 0.2149, "step": 701 }, { "epoch": 0.6512059369202227, "grad_norm": 4.401817798614502, "learning_rate": 6.542627927979772e-06, "loss": 0.2215, "step": 702 }, { "epoch": 0.6521335807050093, "grad_norm": 3.9779934883117676, "learning_rate": 6.5122558567011775e-06, "loss": 0.1763, "step": 703 }, { "epoch": 0.6530612244897959, "grad_norm": 3.7413315773010254, "learning_rate": 6.481920370208274e-06, "loss": 0.1176, "step": 704 }, { "epoch": 0.6539888682745826, "grad_norm": 6.082540988922119, "learning_rate": 6.4516217867059615e-06, "loss": 0.2444, "step": 705 }, { "epoch": 0.6549165120593692, "grad_norm": 7.003627300262451, "learning_rate": 6.421360424012039e-06, "loss": 0.1389, "step": 706 }, { "epoch": 0.6558441558441559, "grad_norm": 4.237902641296387, "learning_rate": 6.39113659955389e-06, "loss": 0.1767, "step": 707 }, { "epoch": 0.6567717996289425, "grad_norm": 3.1200437545776367, "learning_rate": 6.360950630365126e-06, "loss": 0.1127, "step": 708 }, { "epoch": 0.6576994434137291, "grad_norm": 4.659437656402588, "learning_rate": 6.33080283308228e-06, "loss": 0.1292, "step": 709 }, { "epoch": 0.6586270871985158, "grad_norm": 7.017888069152832, "learning_rate": 6.300693523941481e-06, "loss": 0.201, "step": 710 }, { "epoch": 0.6595547309833024, "grad_norm": 4.794656276702881, "learning_rate": 6.270623018775135e-06, "loss": 0.1835, "step": 711 }, { "epoch": 0.660482374768089, "grad_norm": 4.382326602935791, "learning_rate": 6.2405916330086106e-06, "loss": 0.1703, "step": 712 }, { "epoch": 0.6614100185528757, "grad_norm": 6.628244876861572, "learning_rate": 6.210599681656933e-06, "loss": 0.2367, "step": 713 }, { "epoch": 0.6623376623376623, "grad_norm": 4.854303359985352, "learning_rate": 6.180647479321484e-06, "loss": 0.1874, "step": 714 }, { "epoch": 0.6632653061224489, "grad_norm": 2.9809422492980957, "learning_rate": 6.1507353401866896e-06, "loss": 0.1429, "step": 715 }, { "epoch": 0.6641929499072357, "grad_norm": 3.899773359298706, "learning_rate": 6.120863578016736e-06, "loss": 0.1745, "step": 716 }, { "epoch": 0.6651205936920223, "grad_norm": 5.175771713256836, "learning_rate": 6.091032506152274e-06, "loss": 0.162, "step": 717 }, { "epoch": 0.6660482374768089, "grad_norm": 4.2920002937316895, "learning_rate": 6.061242437507131e-06, "loss": 0.1806, "step": 718 }, { "epoch": 0.6669758812615956, "grad_norm": 4.961828708648682, "learning_rate": 6.0314936845650296e-06, "loss": 0.1259, "step": 719 }, { "epoch": 0.6679035250463822, "grad_norm": 4.258510589599609, "learning_rate": 6.00178655937631e-06, "loss": 0.2015, "step": 720 }, { "epoch": 0.6679035250463822, "eval_accuracy": 0.852549889135255, "eval_f1": 0.6825775656324582, "eval_loss": 0.31203413009643555, "eval_precision": 0.8614457831325302, "eval_recall": 0.5652173913043478, "eval_runtime": 47.6328, "eval_samples_per_second": 5.794, "eval_steps_per_second": 0.189, "step": 720 }, { "epoch": 0.6688311688311688, "grad_norm": 4.69655179977417, "learning_rate": 5.972121373554665e-06, "loss": 0.1334, "step": 721 }, { "epoch": 0.6697588126159555, "grad_norm": 4.1556806564331055, "learning_rate": 5.942498438273849e-06, "loss": 0.159, "step": 722 }, { "epoch": 0.6706864564007421, "grad_norm": 4.261734485626221, "learning_rate": 5.912918064264441e-06, "loss": 0.14, "step": 723 }, { "epoch": 0.6716141001855288, "grad_norm": 5.172565460205078, "learning_rate": 5.8833805618105635e-06, "loss": 0.2259, "step": 724 }, { "epoch": 0.6725417439703154, "grad_norm": 5.771614074707031, "learning_rate": 5.853886240746643e-06, "loss": 0.1806, "step": 725 }, { "epoch": 0.673469387755102, "grad_norm": 3.933560371398926, "learning_rate": 5.82443541045415e-06, "loss": 0.1874, "step": 726 }, { "epoch": 0.6743970315398887, "grad_norm": 4.140106201171875, "learning_rate": 5.795028379858355e-06, "loss": 0.1252, "step": 727 }, { "epoch": 0.6753246753246753, "grad_norm": 4.954802513122559, "learning_rate": 5.765665457425102e-06, "loss": 0.2296, "step": 728 }, { "epoch": 0.6762523191094619, "grad_norm": 3.336763620376587, "learning_rate": 5.736346951157544e-06, "loss": 0.1571, "step": 729 }, { "epoch": 0.6771799628942486, "grad_norm": 6.951479434967041, "learning_rate": 5.707073168592943e-06, "loss": 0.227, "step": 730 }, { "epoch": 0.6781076066790352, "grad_norm": 6.168885231018066, "learning_rate": 5.677844416799424e-06, "loss": 0.1793, "step": 731 }, { "epoch": 0.6790352504638218, "grad_norm": 4.202786445617676, "learning_rate": 5.648661002372769e-06, "loss": 0.1307, "step": 732 }, { "epoch": 0.6799628942486086, "grad_norm": 5.333949565887451, "learning_rate": 5.619523231433177e-06, "loss": 0.1435, "step": 733 }, { "epoch": 0.6808905380333952, "grad_norm": 3.218813896179199, "learning_rate": 5.590431409622081e-06, "loss": 0.152, "step": 734 }, { "epoch": 0.6818181818181818, "grad_norm": 3.0740416049957275, "learning_rate": 5.56138584209893e-06, "loss": 0.1625, "step": 735 }, { "epoch": 0.6827458256029685, "grad_norm": 3.7008039951324463, "learning_rate": 5.5323868335379775e-06, "loss": 0.1361, "step": 736 }, { "epoch": 0.6836734693877551, "grad_norm": 3.482449769973755, "learning_rate": 5.503434688125104e-06, "loss": 0.1539, "step": 737 }, { "epoch": 0.6846011131725418, "grad_norm": 3.1356945037841797, "learning_rate": 5.4745297095546125e-06, "loss": 0.116, "step": 738 }, { "epoch": 0.6855287569573284, "grad_norm": 3.0260050296783447, "learning_rate": 5.445672201026054e-06, "loss": 0.1491, "step": 739 }, { "epoch": 0.686456400742115, "grad_norm": 6.386862754821777, "learning_rate": 5.416862465241033e-06, "loss": 0.2216, "step": 740 }, { "epoch": 0.686456400742115, "eval_accuracy": 0.8503325942350333, "eval_f1": 0.6762589928057554, "eval_loss": 0.30298492312431335, "eval_precision": 0.8597560975609756, "eval_recall": 0.5573122529644269, "eval_runtime": 47.363, "eval_samples_per_second": 5.827, "eval_steps_per_second": 0.19, "step": 740 }, { "epoch": 0.6873840445269017, "grad_norm": 2.5552661418914795, "learning_rate": 5.3881008044000495e-06, "loss": 0.113, "step": 741 }, { "epoch": 0.6883116883116883, "grad_norm": 3.5924112796783447, "learning_rate": 5.359387520199317e-06, "loss": 0.0896, "step": 742 }, { "epoch": 0.6892393320964749, "grad_norm": 3.5902557373046875, "learning_rate": 5.330722913827594e-06, "loss": 0.2078, "step": 743 }, { "epoch": 0.6901669758812616, "grad_norm": 3.3769776821136475, "learning_rate": 5.302107285963045e-06, "loss": 0.1311, "step": 744 }, { "epoch": 0.6910946196660482, "grad_norm": 4.20431661605835, "learning_rate": 5.273540936770059e-06, "loss": 0.1694, "step": 745 }, { "epoch": 0.6920222634508348, "grad_norm": 3.680614948272705, "learning_rate": 5.245024165896126e-06, "loss": 0.168, "step": 746 }, { "epoch": 0.6929499072356216, "grad_norm": 4.88540506362915, "learning_rate": 5.216557272468675e-06, "loss": 0.2021, "step": 747 }, { "epoch": 0.6938775510204082, "grad_norm": 3.159914016723633, "learning_rate": 5.18814055509195e-06, "loss": 0.1696, "step": 748 }, { "epoch": 0.6948051948051948, "grad_norm": 5.163516044616699, "learning_rate": 5.1597743118438725e-06, "loss": 0.1485, "step": 749 }, { "epoch": 0.6957328385899815, "grad_norm": 6.143499851226807, "learning_rate": 5.131458840272905e-06, "loss": 0.1769, "step": 750 }, { "epoch": 0.6966604823747681, "grad_norm": 3.032066822052002, "learning_rate": 5.103194437394952e-06, "loss": 0.1202, "step": 751 }, { "epoch": 0.6975881261595547, "grad_norm": 4.9984965324401855, "learning_rate": 5.074981399690219e-06, "loss": 0.2292, "step": 752 }, { "epoch": 0.6985157699443414, "grad_norm": 2.7881083488464355, "learning_rate": 5.046820023100129e-06, "loss": 0.0989, "step": 753 }, { "epoch": 0.699443413729128, "grad_norm": 4.462379455566406, "learning_rate": 5.018710603024187e-06, "loss": 0.1788, "step": 754 }, { "epoch": 0.7003710575139147, "grad_norm": 5.801154613494873, "learning_rate": 4.990653434316915e-06, "loss": 0.1807, "step": 755 }, { "epoch": 0.7012987012987013, "grad_norm": 4.452095031738281, "learning_rate": 4.9626488112847384e-06, "loss": 0.1348, "step": 756 }, { "epoch": 0.7022263450834879, "grad_norm": 4.256060600280762, "learning_rate": 4.934697027682894e-06, "loss": 0.1618, "step": 757 }, { "epoch": 0.7031539888682746, "grad_norm": 2.893348217010498, "learning_rate": 4.9067983767123736e-06, "loss": 0.114, "step": 758 }, { "epoch": 0.7040816326530612, "grad_norm": 2.409883737564087, "learning_rate": 4.878953151016816e-06, "loss": 0.0739, "step": 759 }, { "epoch": 0.7050092764378478, "grad_norm": 4.646303176879883, "learning_rate": 4.851161642679466e-06, "loss": 0.1936, "step": 760 }, { "epoch": 0.7050092764378478, "eval_accuracy": 0.8503325942350333, "eval_f1": 0.6762589928057554, "eval_loss": 0.30913296341896057, "eval_precision": 0.8597560975609756, "eval_recall": 0.5573122529644269, "eval_runtime": 47.2588, "eval_samples_per_second": 5.84, "eval_steps_per_second": 0.19, "step": 760 }, { "epoch": 0.7059369202226345, "grad_norm": 3.7548396587371826, "learning_rate": 4.823424143220097e-06, "loss": 0.1332, "step": 761 }, { "epoch": 0.7068645640074211, "grad_norm": 4.519832611083984, "learning_rate": 4.795740943591955e-06, "loss": 0.1508, "step": 762 }, { "epoch": 0.7077922077922078, "grad_norm": 4.223718643188477, "learning_rate": 4.7681123341787e-06, "loss": 0.2027, "step": 763 }, { "epoch": 0.7087198515769945, "grad_norm": 4.345351219177246, "learning_rate": 4.740538604791371e-06, "loss": 0.1641, "step": 764 }, { "epoch": 0.7096474953617811, "grad_norm": 6.437448024749756, "learning_rate": 4.713020044665348e-06, "loss": 0.1839, "step": 765 }, { "epoch": 0.7105751391465677, "grad_norm": 5.154228687286377, "learning_rate": 4.685556942457296e-06, "loss": 0.1696, "step": 766 }, { "epoch": 0.7115027829313544, "grad_norm": 4.437908172607422, "learning_rate": 4.65814958624217e-06, "loss": 0.1199, "step": 767 }, { "epoch": 0.712430426716141, "grad_norm": 4.337602615356445, "learning_rate": 4.630798263510162e-06, "loss": 0.166, "step": 768 }, { "epoch": 0.7133580705009277, "grad_norm": 4.633522987365723, "learning_rate": 4.60350326116371e-06, "loss": 0.1667, "step": 769 }, { "epoch": 0.7142857142857143, "grad_norm": 4.564393997192383, "learning_rate": 4.576264865514467e-06, "loss": 0.1996, "step": 770 }, { "epoch": 0.7152133580705009, "grad_norm": 5.6954803466796875, "learning_rate": 4.549083362280318e-06, "loss": 0.1227, "step": 771 }, { "epoch": 0.7161410018552876, "grad_norm": 2.704313278198242, "learning_rate": 4.521959036582372e-06, "loss": 0.1118, "step": 772 }, { "epoch": 0.7170686456400742, "grad_norm": 3.529452085494995, "learning_rate": 4.494892172941965e-06, "loss": 0.1595, "step": 773 }, { "epoch": 0.7179962894248608, "grad_norm": 4.921022415161133, "learning_rate": 4.467883055277696e-06, "loss": 0.1755, "step": 774 }, { "epoch": 0.7189239332096475, "grad_norm": 3.468756675720215, "learning_rate": 4.440931966902419e-06, "loss": 0.1198, "step": 775 }, { "epoch": 0.7198515769944341, "grad_norm": 3.8362793922424316, "learning_rate": 4.414039190520308e-06, "loss": 0.162, "step": 776 }, { "epoch": 0.7207792207792207, "grad_norm": 3.9787826538085938, "learning_rate": 4.3872050082238535e-06, "loss": 0.1539, "step": 777 }, { "epoch": 0.7217068645640075, "grad_norm": 4.675662517547607, "learning_rate": 4.360429701490935e-06, "loss": 0.1961, "step": 778 }, { "epoch": 0.7226345083487941, "grad_norm": 3.907136917114258, "learning_rate": 4.3337135511818514e-06, "loss": 0.1566, "step": 779 }, { "epoch": 0.7235621521335807, "grad_norm": 4.647920608520508, "learning_rate": 4.307056837536373e-06, "loss": 0.135, "step": 780 }, { "epoch": 0.7235621521335807, "eval_accuracy": 0.852549889135255, "eval_f1": 0.6855791962174941, "eval_loss": 0.30233433842658997, "eval_precision": 0.8529411764705882, "eval_recall": 0.5731225296442688, "eval_runtime": 49.2875, "eval_samples_per_second": 5.6, "eval_steps_per_second": 0.183, "step": 780 }, { "epoch": 0.7244897959183674, "grad_norm": 4.066539287567139, "learning_rate": 4.280459840170818e-06, "loss": 0.098, "step": 781 }, { "epoch": 0.725417439703154, "grad_norm": 4.152294635772705, "learning_rate": 4.2539228380750955e-06, "loss": 0.1492, "step": 782 }, { "epoch": 0.7263450834879406, "grad_norm": 4.5321831703186035, "learning_rate": 4.2274461096098085e-06, "loss": 0.1727, "step": 783 }, { "epoch": 0.7272727272727273, "grad_norm": 4.524892807006836, "learning_rate": 4.201029932503303e-06, "loss": 0.1529, "step": 784 }, { "epoch": 0.7282003710575139, "grad_norm": 6.373286724090576, "learning_rate": 4.17467458384878e-06, "loss": 0.1686, "step": 785 }, { "epoch": 0.7291280148423006, "grad_norm": 3.8836734294891357, "learning_rate": 4.14838034010138e-06, "loss": 0.1148, "step": 786 }, { "epoch": 0.7300556586270872, "grad_norm": 4.612453937530518, "learning_rate": 4.12214747707527e-06, "loss": 0.1585, "step": 787 }, { "epoch": 0.7309833024118738, "grad_norm": 3.490913152694702, "learning_rate": 4.095976269940777e-06, "loss": 0.1045, "step": 788 }, { "epoch": 0.7319109461966605, "grad_norm": 4.676249027252197, "learning_rate": 4.069866993221473e-06, "loss": 0.2105, "step": 789 }, { "epoch": 0.7328385899814471, "grad_norm": 4.428783416748047, "learning_rate": 4.043819920791322e-06, "loss": 0.1764, "step": 790 }, { "epoch": 0.7337662337662337, "grad_norm": 4.41617488861084, "learning_rate": 4.017835325871781e-06, "loss": 0.1616, "step": 791 }, { "epoch": 0.7346938775510204, "grad_norm": 3.8224010467529297, "learning_rate": 3.991913481028965e-06, "loss": 0.2009, "step": 792 }, { "epoch": 0.735621521335807, "grad_norm": 4.631185531616211, "learning_rate": 3.966054658170754e-06, "loss": 0.1481, "step": 793 }, { "epoch": 0.7365491651205937, "grad_norm": 4.32712984085083, "learning_rate": 3.940259128543967e-06, "loss": 0.1977, "step": 794 }, { "epoch": 0.7374768089053804, "grad_norm": 2.936823606491089, "learning_rate": 3.914527162731498e-06, "loss": 0.1197, "step": 795 }, { "epoch": 0.738404452690167, "grad_norm": 6.375387191772461, "learning_rate": 3.888859030649498e-06, "loss": 0.2587, "step": 796 }, { "epoch": 0.7393320964749536, "grad_norm": 5.587421894073486, "learning_rate": 3.863255001544526e-06, "loss": 0.2271, "step": 797 }, { "epoch": 0.7402597402597403, "grad_norm": 5.898324966430664, "learning_rate": 3.837715343990727e-06, "loss": 0.2026, "step": 798 }, { "epoch": 0.7411873840445269, "grad_norm": 6.89965295791626, "learning_rate": 3.81224032588703e-06, "loss": 0.222, "step": 799 }, { "epoch": 0.7421150278293135, "grad_norm": 4.845470428466797, "learning_rate": 3.7868302144543146e-06, "loss": 0.1332, "step": 800 }, { "epoch": 0.7421150278293135, "eval_accuracy": 0.843680709534368, "eval_f1": 0.6466165413533834, "eval_loss": 0.32069557905197144, "eval_precision": 0.8835616438356164, "eval_recall": 0.5098814229249012, "eval_runtime": 47.5467, "eval_samples_per_second": 5.805, "eval_steps_per_second": 0.189, "step": 800 }, { "epoch": 0.7430426716141002, "grad_norm": 3.8346760272979736, "learning_rate": 3.7614852762326303e-06, "loss": 0.1511, "step": 801 }, { "epoch": 0.7439703153988868, "grad_norm": 3.3582327365875244, "learning_rate": 3.736205777078381e-06, "loss": 0.099, "step": 802 }, { "epoch": 0.7448979591836735, "grad_norm": 4.043883323669434, "learning_rate": 3.7109919821615546e-06, "loss": 0.1684, "step": 803 }, { "epoch": 0.7458256029684601, "grad_norm": 4.061108112335205, "learning_rate": 3.685844155962931e-06, "loss": 0.1874, "step": 804 }, { "epoch": 0.7467532467532467, "grad_norm": 2.596107244491577, "learning_rate": 3.6607625622713005e-06, "loss": 0.1146, "step": 805 }, { "epoch": 0.7476808905380334, "grad_norm": 3.7484891414642334, "learning_rate": 3.63574746418072e-06, "loss": 0.1027, "step": 806 }, { "epoch": 0.74860853432282, "grad_norm": 4.787120342254639, "learning_rate": 3.610799124087725e-06, "loss": 0.2284, "step": 807 }, { "epoch": 0.7495361781076066, "grad_norm": 3.2740108966827393, "learning_rate": 3.585917803688603e-06, "loss": 0.126, "step": 808 }, { "epoch": 0.7504638218923934, "grad_norm": 3.000568389892578, "learning_rate": 3.5611037639766267e-06, "loss": 0.1258, "step": 809 }, { "epoch": 0.75139146567718, "grad_norm": 3.994319200515747, "learning_rate": 3.536357265239333e-06, "loss": 0.2042, "step": 810 }, { "epoch": 0.7523191094619666, "grad_norm": 3.8152012825012207, "learning_rate": 3.511678567055786e-06, "loss": 0.147, "step": 811 }, { "epoch": 0.7532467532467533, "grad_norm": 3.8970158100128174, "learning_rate": 3.487067928293848e-06, "loss": 0.142, "step": 812 }, { "epoch": 0.7541743970315399, "grad_norm": 4.383775234222412, "learning_rate": 3.4625256071074776e-06, "loss": 0.1752, "step": 813 }, { "epoch": 0.7551020408163265, "grad_norm": 3.43391489982605, "learning_rate": 3.4380518609340076e-06, "loss": 0.1207, "step": 814 }, { "epoch": 0.7560296846011132, "grad_norm": 3.767289161682129, "learning_rate": 3.413646946491458e-06, "loss": 0.1301, "step": 815 }, { "epoch": 0.7569573283858998, "grad_norm": 3.6454761028289795, "learning_rate": 3.3893111197758276e-06, "loss": 0.1729, "step": 816 }, { "epoch": 0.7578849721706865, "grad_norm": 4.36724328994751, "learning_rate": 3.3650446360584276e-06, "loss": 0.1707, "step": 817 }, { "epoch": 0.7588126159554731, "grad_norm": 2.9927897453308105, "learning_rate": 3.3408477498831917e-06, "loss": 0.112, "step": 818 }, { "epoch": 0.7597402597402597, "grad_norm": 4.448176860809326, "learning_rate": 3.3167207150640003e-06, "loss": 0.1064, "step": 819 }, { "epoch": 0.7606679035250464, "grad_norm": 6.467900276184082, "learning_rate": 3.2926637846820366e-06, "loss": 0.249, "step": 820 }, { "epoch": 0.7606679035250464, "eval_accuracy": 0.8592017738359202, "eval_f1": 0.6997635933806147, "eval_loss": 0.3030702769756317, "eval_precision": 0.8705882352941177, "eval_recall": 0.5849802371541502, "eval_runtime": 47.6515, "eval_samples_per_second": 5.792, "eval_steps_per_second": 0.189, "step": 820 }, { "epoch": 0.761595547309833, "grad_norm": 3.21474027633667, "learning_rate": 3.268677211083109e-06, "loss": 0.1103, "step": 821 }, { "epoch": 0.7625231910946196, "grad_norm": 3.8230926990509033, "learning_rate": 3.2447612458750365e-06, "loss": 0.1608, "step": 822 }, { "epoch": 0.7634508348794063, "grad_norm": 4.736266136169434, "learning_rate": 3.2209161399249677e-06, "loss": 0.1549, "step": 823 }, { "epoch": 0.764378478664193, "grad_norm": 5.4161481857299805, "learning_rate": 3.197142143356787e-06, "loss": 0.1905, "step": 824 }, { "epoch": 0.7653061224489796, "grad_norm": 3.857360601425171, "learning_rate": 3.1734395055484623e-06, "loss": 0.1287, "step": 825 }, { "epoch": 0.7662337662337663, "grad_norm": 4.246245861053467, "learning_rate": 3.1498084751294523e-06, "loss": 0.1751, "step": 826 }, { "epoch": 0.7671614100185529, "grad_norm": 4.21674108505249, "learning_rate": 3.126249299978086e-06, "loss": 0.1593, "step": 827 }, { "epoch": 0.7680890538033395, "grad_norm": 3.7095324993133545, "learning_rate": 3.1027622272189572e-06, "loss": 0.1384, "step": 828 }, { "epoch": 0.7690166975881262, "grad_norm": 6.3794965744018555, "learning_rate": 3.0793475032203513e-06, "loss": 0.1583, "step": 829 }, { "epoch": 0.7699443413729128, "grad_norm": 3.0277578830718994, "learning_rate": 3.0560053735916372e-06, "loss": 0.1043, "step": 830 }, { "epoch": 0.7708719851576994, "grad_norm": 5.117831707000732, "learning_rate": 3.032736083180716e-06, "loss": 0.15, "step": 831 }, { "epoch": 0.7717996289424861, "grad_norm": 2.76505184173584, "learning_rate": 3.009539876071427e-06, "loss": 0.0558, "step": 832 }, { "epoch": 0.7727272727272727, "grad_norm": 4.057276725769043, "learning_rate": 2.9864169955810085e-06, "loss": 0.1391, "step": 833 }, { "epoch": 0.7736549165120594, "grad_norm": 3.700852870941162, "learning_rate": 2.9633676842575386e-06, "loss": 0.1721, "step": 834 }, { "epoch": 0.774582560296846, "grad_norm": 4.1468939781188965, "learning_rate": 2.940392183877382e-06, "loss": 0.1866, "step": 835 }, { "epoch": 0.7755102040816326, "grad_norm": 3.1190547943115234, "learning_rate": 2.9174907354426696e-06, "loss": 0.1292, "step": 836 }, { "epoch": 0.7764378478664193, "grad_norm": 3.0235095024108887, "learning_rate": 2.8946635791787546e-06, "loss": 0.0629, "step": 837 }, { "epoch": 0.7773654916512059, "grad_norm": 3.435035467147827, "learning_rate": 2.8719109545317102e-06, "loss": 0.1064, "step": 838 }, { "epoch": 0.7782931354359925, "grad_norm": 5.368072986602783, "learning_rate": 2.849233100165795e-06, "loss": 0.1662, "step": 839 }, { "epoch": 0.7792207792207793, "grad_norm": 4.488304615020752, "learning_rate": 2.8266302539609747e-06, "loss": 0.2033, "step": 840 }, { "epoch": 0.7792207792207793, "eval_accuracy": 0.8592017738359202, "eval_f1": 0.6983372921615202, "eval_loss": 0.30760514736175537, "eval_precision": 0.875, "eval_recall": 0.5810276679841897, "eval_runtime": 47.6262, "eval_samples_per_second": 5.795, "eval_steps_per_second": 0.189, "step": 840 }, { "epoch": 0.7801484230055659, "grad_norm": 4.42378044128418, "learning_rate": 2.8041026530104144e-06, "loss": 0.1223, "step": 841 }, { "epoch": 0.7810760667903525, "grad_norm": 3.143146514892578, "learning_rate": 2.78165053361798e-06, "loss": 0.1408, "step": 842 }, { "epoch": 0.7820037105751392, "grad_norm": 3.8427817821502686, "learning_rate": 2.759274131295787e-06, "loss": 0.0995, "step": 843 }, { "epoch": 0.7829313543599258, "grad_norm": 4.896306991577148, "learning_rate": 2.736973680761702e-06, "loss": 0.1621, "step": 844 }, { "epoch": 0.7838589981447124, "grad_norm": 5.344152450561523, "learning_rate": 2.714749415936904e-06, "loss": 0.1749, "step": 845 }, { "epoch": 0.7847866419294991, "grad_norm": 4.093927383422852, "learning_rate": 2.692601569943407e-06, "loss": 0.1716, "step": 846 }, { "epoch": 0.7857142857142857, "grad_norm": 3.3568267822265625, "learning_rate": 2.670530375101641e-06, "loss": 0.1658, "step": 847 }, { "epoch": 0.7866419294990723, "grad_norm": 3.3225290775299072, "learning_rate": 2.648536062927999e-06, "loss": 0.1358, "step": 848 }, { "epoch": 0.787569573283859, "grad_norm": 4.490353584289551, "learning_rate": 2.6266188641324e-06, "loss": 0.1178, "step": 849 }, { "epoch": 0.7884972170686456, "grad_norm": 3.4706945419311523, "learning_rate": 2.604779008615895e-06, "loss": 0.0946, "step": 850 }, { "epoch": 0.7894248608534323, "grad_norm": 5.027894020080566, "learning_rate": 2.583016725468226e-06, "loss": 0.1434, "step": 851 }, { "epoch": 0.7903525046382189, "grad_norm": 4.010559558868408, "learning_rate": 2.5613322429654573e-06, "loss": 0.158, "step": 852 }, { "epoch": 0.7912801484230055, "grad_norm": 6.057810306549072, "learning_rate": 2.5397257885675396e-06, "loss": 0.1912, "step": 853 }, { "epoch": 0.7922077922077922, "grad_norm": 5.5663371086120605, "learning_rate": 2.5181975889159615e-06, "loss": 0.2398, "step": 854 }, { "epoch": 0.7931354359925789, "grad_norm": 2.8672025203704834, "learning_rate": 2.496747869831345e-06, "loss": 0.0944, "step": 855 }, { "epoch": 0.7940630797773655, "grad_norm": 4.048581123352051, "learning_rate": 2.475376856311097e-06, "loss": 0.1603, "step": 856 }, { "epoch": 0.7949907235621522, "grad_norm": 5.737659931182861, "learning_rate": 2.4540847725270376e-06, "loss": 0.1366, "step": 857 }, { "epoch": 0.7959183673469388, "grad_norm": 4.4261155128479, "learning_rate": 2.432871841823047e-06, "loss": 0.1939, "step": 858 }, { "epoch": 0.7968460111317254, "grad_norm": 5.629834175109863, "learning_rate": 2.411738286712735e-06, "loss": 0.2281, "step": 859 }, { "epoch": 0.7977736549165121, "grad_norm": 3.919034004211426, "learning_rate": 2.390684328877089e-06, "loss": 0.1418, "step": 860 }, { "epoch": 0.7977736549165121, "eval_accuracy": 0.8614190687361419, "eval_f1": 0.7072599531615925, "eval_loss": 0.29977986216545105, "eval_precision": 0.867816091954023, "eval_recall": 0.5968379446640316, "eval_runtime": 47.4113, "eval_samples_per_second": 5.821, "eval_steps_per_second": 0.19, "step": 860 }, { "epoch": 0.7987012987012987, "grad_norm": 7.607851982116699, "learning_rate": 2.36971018916217e-06, "loss": 0.187, "step": 861 }, { "epoch": 0.7996289424860853, "grad_norm": 3.1179118156433105, "learning_rate": 2.3488160875767717e-06, "loss": 0.1326, "step": 862 }, { "epoch": 0.800556586270872, "grad_norm": 3.8754749298095703, "learning_rate": 2.328002243290138e-06, "loss": 0.1497, "step": 863 }, { "epoch": 0.8014842300556586, "grad_norm": 4.216552734375, "learning_rate": 2.307268874629649e-06, "loss": 0.1259, "step": 864 }, { "epoch": 0.8024118738404453, "grad_norm": 5.980984210968018, "learning_rate": 2.2866161990785228e-06, "loss": 0.1778, "step": 865 }, { "epoch": 0.8033395176252319, "grad_norm": 4.638891220092773, "learning_rate": 2.266044433273562e-06, "loss": 0.2258, "step": 866 }, { "epoch": 0.8042671614100185, "grad_norm": 3.3244616985321045, "learning_rate": 2.245553793002849e-06, "loss": 0.1215, "step": 867 }, { "epoch": 0.8051948051948052, "grad_norm": 6.757506370544434, "learning_rate": 2.2251444932035094e-06, "loss": 0.159, "step": 868 }, { "epoch": 0.8061224489795918, "grad_norm": 5.527317523956299, "learning_rate": 2.204816747959434e-06, "loss": 0.1025, "step": 869 }, { "epoch": 0.8070500927643784, "grad_norm": 3.5259664058685303, "learning_rate": 2.184570770499056e-06, "loss": 0.1522, "step": 870 }, { "epoch": 0.8079777365491652, "grad_norm": 4.936224937438965, "learning_rate": 2.1644067731931005e-06, "loss": 0.1679, "step": 871 }, { "epoch": 0.8089053803339518, "grad_norm": 4.338299751281738, "learning_rate": 2.1443249675523536e-06, "loss": 0.1705, "step": 872 }, { "epoch": 0.8098330241187384, "grad_norm": 3.631812334060669, "learning_rate": 2.124325564225458e-06, "loss": 0.1034, "step": 873 }, { "epoch": 0.8107606679035251, "grad_norm": 3.276409864425659, "learning_rate": 2.1044087729966856e-06, "loss": 0.1486, "step": 874 }, { "epoch": 0.8116883116883117, "grad_norm": 5.677032947540283, "learning_rate": 2.0845748027837585e-06, "loss": 0.2155, "step": 875 }, { "epoch": 0.8126159554730983, "grad_norm": 6.064774036407471, "learning_rate": 2.064823861635633e-06, "loss": 0.1773, "step": 876 }, { "epoch": 0.813543599257885, "grad_norm": 3.783052444458008, "learning_rate": 2.0451561567303378e-06, "loss": 0.1696, "step": 877 }, { "epoch": 0.8144712430426716, "grad_norm": 3.9778928756713867, "learning_rate": 2.025571894372794e-06, "loss": 0.0916, "step": 878 }, { "epoch": 0.8153988868274582, "grad_norm": 3.2632224559783936, "learning_rate": 2.0060712799926407e-06, "loss": 0.1004, "step": 879 }, { "epoch": 0.8163265306122449, "grad_norm": 5.624824047088623, "learning_rate": 1.9866545181421016e-06, "loss": 0.1826, "step": 880 }, { "epoch": 0.8163265306122449, "eval_accuracy": 0.8625277161862528, "eval_f1": 0.7089201877934272, "eval_loss": 0.30142825841903687, "eval_precision": 0.8728323699421965, "eval_recall": 0.5968379446640316, "eval_runtime": 47.5017, "eval_samples_per_second": 5.81, "eval_steps_per_second": 0.189, "step": 880 }, { "epoch": 0.8172541743970315, "grad_norm": 5.620955944061279, "learning_rate": 1.967321812493813e-06, "loss": 0.159, "step": 881 }, { "epoch": 0.8181818181818182, "grad_norm": 3.5917294025421143, "learning_rate": 1.9480733658387175e-06, "loss": 0.1544, "step": 882 }, { "epoch": 0.8191094619666048, "grad_norm": 4.478516101837158, "learning_rate": 1.9289093800839067e-06, "loss": 0.1835, "step": 883 }, { "epoch": 0.8200371057513914, "grad_norm": 5.72329044342041, "learning_rate": 1.9098300562505266e-06, "loss": 0.181, "step": 884 }, { "epoch": 0.8209647495361782, "grad_norm": 4.014090538024902, "learning_rate": 1.8908355944716516e-06, "loss": 0.123, "step": 885 }, { "epoch": 0.8218923933209648, "grad_norm": 3.9502549171447754, "learning_rate": 1.8719261939902023e-06, "loss": 0.1675, "step": 886 }, { "epoch": 0.8228200371057514, "grad_norm": 3.701931953430176, "learning_rate": 1.8531020531568377e-06, "loss": 0.1297, "step": 887 }, { "epoch": 0.8237476808905381, "grad_norm": 4.026066303253174, "learning_rate": 1.8343633694278895e-06, "loss": 0.1366, "step": 888 }, { "epoch": 0.8246753246753247, "grad_norm": 4.122823715209961, "learning_rate": 1.8157103393632869e-06, "loss": 0.1819, "step": 889 }, { "epoch": 0.8256029684601113, "grad_norm": 4.512097358703613, "learning_rate": 1.7971431586244814e-06, "loss": 0.2281, "step": 890 }, { "epoch": 0.826530612244898, "grad_norm": 3.6927201747894287, "learning_rate": 1.7786620219724205e-06, "loss": 0.0749, "step": 891 }, { "epoch": 0.8274582560296846, "grad_norm": 5.3355207443237305, "learning_rate": 1.7602671232654755e-06, "loss": 0.1781, "step": 892 }, { "epoch": 0.8283858998144712, "grad_norm": 3.310504674911499, "learning_rate": 1.7419586554574364e-06, "loss": 0.0927, "step": 893 }, { "epoch": 0.8293135435992579, "grad_norm": 3.6718716621398926, "learning_rate": 1.723736810595461e-06, "loss": 0.159, "step": 894 }, { "epoch": 0.8302411873840445, "grad_norm": 4.083915710449219, "learning_rate": 1.7056017798180824e-06, "loss": 0.1516, "step": 895 }, { "epoch": 0.8311688311688312, "grad_norm": 8.39147663116455, "learning_rate": 1.687553753353195e-06, "loss": 0.1548, "step": 896 }, { "epoch": 0.8320964749536178, "grad_norm": 3.83030366897583, "learning_rate": 1.669592920516049e-06, "loss": 0.1613, "step": 897 }, { "epoch": 0.8330241187384044, "grad_norm": 3.559238910675049, "learning_rate": 1.6517194697072903e-06, "loss": 0.1193, "step": 898 }, { "epoch": 0.8339517625231911, "grad_norm": 4.015861511230469, "learning_rate": 1.633933588410952e-06, "loss": 0.1383, "step": 899 }, { "epoch": 0.8348794063079777, "grad_norm": 4.48312520980835, "learning_rate": 1.6162354631925203e-06, "loss": 0.1538, "step": 900 }, { "epoch": 0.8348794063079777, "eval_accuracy": 0.8614190687361419, "eval_f1": 0.7016706443914081, "eval_loss": 0.3091951012611389, "eval_precision": 0.8855421686746988, "eval_recall": 0.5810276679841897, "eval_runtime": 47.4361, "eval_samples_per_second": 5.818, "eval_steps_per_second": 0.19, "step": 900 }, { "epoch": 0.8358070500927643, "grad_norm": 3.852740526199341, "learning_rate": 1.5986252796969482e-06, "loss": 0.1165, "step": 901 }, { "epoch": 0.8367346938775511, "grad_norm": 5.131833076477051, "learning_rate": 1.5811032226467304e-06, "loss": 0.198, "step": 902 }, { "epoch": 0.8376623376623377, "grad_norm": 4.975651741027832, "learning_rate": 1.5636694758399563e-06, "loss": 0.1891, "step": 903 }, { "epoch": 0.8385899814471243, "grad_norm": 3.24419903755188, "learning_rate": 1.5463242221483742e-06, "loss": 0.0935, "step": 904 }, { "epoch": 0.839517625231911, "grad_norm": 3.5641651153564453, "learning_rate": 1.5290676435154949e-06, "loss": 0.1533, "step": 905 }, { "epoch": 0.8404452690166976, "grad_norm": 3.872134208679199, "learning_rate": 1.511899920954656e-06, "loss": 0.1545, "step": 906 }, { "epoch": 0.8413729128014842, "grad_norm": 6.075543403625488, "learning_rate": 1.4948212345471492e-06, "loss": 0.2032, "step": 907 }, { "epoch": 0.8423005565862709, "grad_norm": 2.9056954383850098, "learning_rate": 1.4778317634403082e-06, "loss": 0.0986, "step": 908 }, { "epoch": 0.8432282003710575, "grad_norm": 5.516162872314453, "learning_rate": 1.460931685845649e-06, "loss": 0.1868, "step": 909 }, { "epoch": 0.8441558441558441, "grad_norm": 2.8610849380493164, "learning_rate": 1.4441211790369892e-06, "loss": 0.0923, "step": 910 }, { "epoch": 0.8450834879406308, "grad_norm": 3.700863838195801, "learning_rate": 1.427400419348588e-06, "loss": 0.1291, "step": 911 }, { "epoch": 0.8460111317254174, "grad_norm": 4.772455215454102, "learning_rate": 1.4107695821733026e-06, "loss": 0.1352, "step": 912 }, { "epoch": 0.8469387755102041, "grad_norm": 3.5742745399475098, "learning_rate": 1.3942288419607476e-06, "loss": 0.1824, "step": 913 }, { "epoch": 0.8478664192949907, "grad_norm": 8.259415626525879, "learning_rate": 1.3777783722154603e-06, "loss": 0.2448, "step": 914 }, { "epoch": 0.8487940630797773, "grad_norm": 3.900238513946533, "learning_rate": 1.3614183454950824e-06, "loss": 0.1273, "step": 915 }, { "epoch": 0.849721706864564, "grad_norm": 2.9773433208465576, "learning_rate": 1.3451489334085555e-06, "loss": 0.1522, "step": 916 }, { "epoch": 0.8506493506493507, "grad_norm": 3.071232318878174, "learning_rate": 1.3289703066143112e-06, "loss": 0.1256, "step": 917 }, { "epoch": 0.8515769944341373, "grad_norm": 3.8165667057037354, "learning_rate": 1.3128826348184886e-06, "loss": 0.1111, "step": 918 }, { "epoch": 0.852504638218924, "grad_norm": 3.7821688652038574, "learning_rate": 1.296886086773157e-06, "loss": 0.2091, "step": 919 }, { "epoch": 0.8534322820037106, "grad_norm": 4.833895206451416, "learning_rate": 1.2809808302745298e-06, "loss": 0.1762, "step": 920 }, { "epoch": 0.8534322820037106, "eval_accuracy": 0.8603104212860311, "eval_f1": 0.704225352112676, "eval_loss": 0.30113720893859863, "eval_precision": 0.8670520231213873, "eval_recall": 0.5928853754940712, "eval_runtime": 47.313, "eval_samples_per_second": 5.833, "eval_steps_per_second": 0.19, "step": 920 }, { "epoch": 0.8543599257884972, "grad_norm": 3.3207972049713135, "learning_rate": 1.2651670321612264e-06, "loss": 0.1367, "step": 921 }, { "epoch": 0.8552875695732839, "grad_norm": 3.202796697616577, "learning_rate": 1.249444858312502e-06, "loss": 0.1379, "step": 922 }, { "epoch": 0.8562152133580705, "grad_norm": 6.188356876373291, "learning_rate": 1.233814473646524e-06, "loss": 0.2627, "step": 923 }, { "epoch": 0.8571428571428571, "grad_norm": 3.4624321460723877, "learning_rate": 1.218276042118629e-06, "loss": 0.1318, "step": 924 }, { "epoch": 0.8580705009276438, "grad_norm": 3.288809061050415, "learning_rate": 1.202829726719611e-06, "loss": 0.1188, "step": 925 }, { "epoch": 0.8589981447124304, "grad_norm": 2.691675901412964, "learning_rate": 1.1874756894740137e-06, "loss": 0.1252, "step": 926 }, { "epoch": 0.859925788497217, "grad_norm": 3.750600576400757, "learning_rate": 1.1722140914384162e-06, "loss": 0.1644, "step": 927 }, { "epoch": 0.8608534322820037, "grad_norm": 3.1353397369384766, "learning_rate": 1.1570450926997657e-06, "loss": 0.1461, "step": 928 }, { "epoch": 0.8617810760667903, "grad_norm": 5.295469760894775, "learning_rate": 1.1419688523736761e-06, "loss": 0.1967, "step": 929 }, { "epoch": 0.862708719851577, "grad_norm": 3.461599349975586, "learning_rate": 1.1269855286027798e-06, "loss": 0.1426, "step": 930 }, { "epoch": 0.8636363636363636, "grad_norm": 6.9660420417785645, "learning_rate": 1.1120952785550477e-06, "loss": 0.2015, "step": 931 }, { "epoch": 0.8645640074211502, "grad_norm": 2.989213705062866, "learning_rate": 1.0972982584221592e-06, "loss": 0.1204, "step": 932 }, { "epoch": 0.865491651205937, "grad_norm": 4.492414474487305, "learning_rate": 1.0825946234178575e-06, "loss": 0.1579, "step": 933 }, { "epoch": 0.8664192949907236, "grad_norm": 4.693439960479736, "learning_rate": 1.067984527776309e-06, "loss": 0.1959, "step": 934 }, { "epoch": 0.8673469387755102, "grad_norm": 5.462426662445068, "learning_rate": 1.0534681247505107e-06, "loss": 0.1435, "step": 935 }, { "epoch": 0.8682745825602969, "grad_norm": 2.594604730606079, "learning_rate": 1.0390455666106547e-06, "loss": 0.115, "step": 936 }, { "epoch": 0.8692022263450835, "grad_norm": 5.900606155395508, "learning_rate": 1.024717004642557e-06, "loss": 0.1749, "step": 937 }, { "epoch": 0.8701298701298701, "grad_norm": 5.774359226226807, "learning_rate": 1.010482589146048e-06, "loss": 0.1802, "step": 938 }, { "epoch": 0.8710575139146568, "grad_norm": 4.002913951873779, "learning_rate": 9.963424694334122e-07, "loss": 0.1277, "step": 939 }, { "epoch": 0.8719851576994434, "grad_norm": 3.6173672676086426, "learning_rate": 9.822967938278172e-07, "loss": 0.1561, "step": 940 }, { "epoch": 0.8719851576994434, "eval_accuracy": 0.8603104212860311, "eval_f1": 0.704225352112676, "eval_loss": 0.29984721541404724, "eval_precision": 0.8670520231213873, "eval_recall": 0.5928853754940712, "eval_runtime": 48.0345, "eval_samples_per_second": 5.746, "eval_steps_per_second": 0.187, "step": 940 }, { "epoch": 0.87291280148423, "grad_norm": 5.298496723175049, "learning_rate": 9.683457096617487e-07, "loss": 0.1343, "step": 941 }, { "epoch": 0.8738404452690167, "grad_norm": 4.087591648101807, "learning_rate": 9.544893632754816e-07, "loss": 0.1342, "step": 942 }, { "epoch": 0.8747680890538033, "grad_norm": 3.6953861713409424, "learning_rate": 9.407279000155311e-07, "loss": 0.1125, "step": 943 }, { "epoch": 0.87569573283859, "grad_norm": 5.693349838256836, "learning_rate": 9.270614642331377e-07, "loss": 0.2285, "step": 944 }, { "epoch": 0.8766233766233766, "grad_norm": 4.321276664733887, "learning_rate": 9.134901992827427e-07, "loss": 0.2169, "step": 945 }, { "epoch": 0.8775510204081632, "grad_norm": 5.951560020446777, "learning_rate": 9.000142475204965e-07, "loss": 0.2039, "step": 946 }, { "epoch": 0.87847866419295, "grad_norm": 5.382765293121338, "learning_rate": 8.866337503027523e-07, "loss": 0.1347, "step": 947 }, { "epoch": 0.8794063079777366, "grad_norm": 4.566171646118164, "learning_rate": 8.733488479845997e-07, "loss": 0.1929, "step": 948 }, { "epoch": 0.8803339517625232, "grad_norm": 4.413459300994873, "learning_rate": 8.60159679918372e-07, "loss": 0.1463, "step": 949 }, { "epoch": 0.8812615955473099, "grad_norm": 3.8674092292785645, "learning_rate": 8.470663844522053e-07, "loss": 0.1523, "step": 950 }, { "epoch": 0.8821892393320965, "grad_norm": 3.844576597213745, "learning_rate": 8.340690989285727e-07, "loss": 0.1248, "step": 951 }, { "epoch": 0.8831168831168831, "grad_norm": 4.541808605194092, "learning_rate": 8.211679596828481e-07, "loss": 0.1571, "step": 952 }, { "epoch": 0.8840445269016698, "grad_norm": 3.0702145099639893, "learning_rate": 8.083631020418792e-07, "loss": 0.157, "step": 953 }, { "epoch": 0.8849721706864564, "grad_norm": 3.5125439167022705, "learning_rate": 7.956546603225601e-07, "loss": 0.1011, "step": 954 }, { "epoch": 0.885899814471243, "grad_norm": 4.256104469299316, "learning_rate": 7.830427678304353e-07, "loss": 0.1411, "step": 955 }, { "epoch": 0.8868274582560297, "grad_norm": 4.931686878204346, "learning_rate": 7.705275568582848e-07, "loss": 0.1953, "step": 956 }, { "epoch": 0.8877551020408163, "grad_norm": 5.233354091644287, "learning_rate": 7.581091586847522e-07, "loss": 0.2095, "step": 957 }, { "epoch": 0.8886827458256029, "grad_norm": 6.383068084716797, "learning_rate": 7.457877035729588e-07, "loss": 0.2274, "step": 958 }, { "epoch": 0.8896103896103896, "grad_norm": 2.8475682735443115, "learning_rate": 7.335633207691362e-07, "loss": 0.1336, "step": 959 }, { "epoch": 0.8905380333951762, "grad_norm": 3.393915891647339, "learning_rate": 7.21436138501278e-07, "loss": 0.1633, "step": 960 }, { "epoch": 0.8905380333951762, "eval_accuracy": 0.8569844789356984, "eval_f1": 0.6935866983372921, "eval_loss": 0.3064272701740265, "eval_precision": 0.8690476190476191, "eval_recall": 0.5770750988142292, "eval_runtime": 48.2701, "eval_samples_per_second": 5.718, "eval_steps_per_second": 0.186, "step": 960 }, { "epoch": 0.891465677179963, "grad_norm": 4.68550968170166, "learning_rate": 7.094062839777838e-07, "loss": 0.1854, "step": 961 }, { "epoch": 0.8923933209647495, "grad_norm": 5.072958946228027, "learning_rate": 6.974738833861383e-07, "loss": 0.1762, "step": 962 }, { "epoch": 0.8933209647495362, "grad_norm": 4.519327640533447, "learning_rate": 6.856390618915775e-07, "loss": 0.182, "step": 963 }, { "epoch": 0.8942486085343229, "grad_norm": 5.558988094329834, "learning_rate": 6.739019436357774e-07, "loss": 0.1665, "step": 964 }, { "epoch": 0.8951762523191095, "grad_norm": 2.263278007507324, "learning_rate": 6.622626517355557e-07, "loss": 0.1112, "step": 965 }, { "epoch": 0.8961038961038961, "grad_norm": 5.888603687286377, "learning_rate": 6.507213082815745e-07, "loss": 0.1455, "step": 966 }, { "epoch": 0.8970315398886828, "grad_norm": 5.091086387634277, "learning_rate": 6.392780343370686e-07, "loss": 0.1812, "step": 967 }, { "epoch": 0.8979591836734694, "grad_norm": 6.290548324584961, "learning_rate": 6.279329499365649e-07, "loss": 0.1527, "step": 968 }, { "epoch": 0.898886827458256, "grad_norm": 6.533473014831543, "learning_rate": 6.166861740846297e-07, "loss": 0.2105, "step": 969 }, { "epoch": 0.8998144712430427, "grad_norm": 3.4495279788970947, "learning_rate": 6.055378247546217e-07, "loss": 0.1222, "step": 970 }, { "epoch": 0.9007421150278293, "grad_norm": 5.290384769439697, "learning_rate": 5.94488018887448e-07, "loss": 0.2046, "step": 971 }, { "epoch": 0.9016697588126159, "grad_norm": 6.091614723205566, "learning_rate": 5.835368723903456e-07, "loss": 0.2643, "step": 972 }, { "epoch": 0.9025974025974026, "grad_norm": 4.488548278808594, "learning_rate": 5.726845001356573e-07, "loss": 0.1263, "step": 973 }, { "epoch": 0.9035250463821892, "grad_norm": 2.7875099182128906, "learning_rate": 5.619310159596358e-07, "loss": 0.0922, "step": 974 }, { "epoch": 0.9044526901669759, "grad_norm": 5.558516025543213, "learning_rate": 5.51276532661238e-07, "loss": 0.2045, "step": 975 }, { "epoch": 0.9053803339517625, "grad_norm": 5.901011943817139, "learning_rate": 5.407211620009545e-07, "loss": 0.1743, "step": 976 }, { "epoch": 0.9063079777365491, "grad_norm": 3.838674783706665, "learning_rate": 5.30265014699628e-07, "loss": 0.1728, "step": 977 }, { "epoch": 0.9072356215213359, "grad_norm": 3.811453104019165, "learning_rate": 5.199082004372958e-07, "loss": 0.153, "step": 978 }, { "epoch": 0.9081632653061225, "grad_norm": 5.14892578125, "learning_rate": 5.096508278520385e-07, "loss": 0.1991, "step": 979 }, { "epoch": 0.9090909090909091, "grad_norm": 3.6292712688446045, "learning_rate": 4.994930045388414e-07, "loss": 0.1452, "step": 980 }, { "epoch": 0.9090909090909091, "eval_accuracy": 0.8603104212860311, "eval_f1": 0.7028301886792453, "eval_loss": 0.3034472167491913, "eval_precision": 0.8713450292397661, "eval_recall": 0.5889328063241107, "eval_runtime": 48.7572, "eval_samples_per_second": 5.661, "eval_steps_per_second": 0.185, "step": 980 }, { "epoch": 0.9100185528756958, "grad_norm": 3.438109874725342, "learning_rate": 4.894348370484648e-07, "loss": 0.1054, "step": 981 }, { "epoch": 0.9109461966604824, "grad_norm": 5.481462478637695, "learning_rate": 4.794764308863242e-07, "loss": 0.1463, "step": 982 }, { "epoch": 0.911873840445269, "grad_norm": 6.784456253051758, "learning_rate": 4.696178905113913e-07, "loss": 0.1634, "step": 983 }, { "epoch": 0.9128014842300557, "grad_norm": 3.902355194091797, "learning_rate": 4.5985931933508757e-07, "loss": 0.1689, "step": 984 }, { "epoch": 0.9137291280148423, "grad_norm": 4.524623394012451, "learning_rate": 4.502008197202068e-07, "loss": 0.1428, "step": 985 }, { "epoch": 0.9146567717996289, "grad_norm": 3.976349353790283, "learning_rate": 4.406424929798403e-07, "loss": 0.1864, "step": 986 }, { "epoch": 0.9155844155844156, "grad_norm": 3.3905527591705322, "learning_rate": 4.3118443937631094e-07, "loss": 0.1719, "step": 987 }, { "epoch": 0.9165120593692022, "grad_norm": 4.1316938400268555, "learning_rate": 4.218267581201296e-07, "loss": 0.1124, "step": 988 }, { "epoch": 0.9174397031539888, "grad_norm": 5.6381754875183105, "learning_rate": 4.125695473689406e-07, "loss": 0.1994, "step": 989 }, { "epoch": 0.9183673469387755, "grad_norm": 7.625948905944824, "learning_rate": 4.034129042265067e-07, "loss": 0.2211, "step": 990 }, { "epoch": 0.9192949907235621, "grad_norm": 3.567246437072754, "learning_rate": 3.943569247416801e-07, "loss": 0.1359, "step": 991 }, { "epoch": 0.9202226345083488, "grad_norm": 4.336119174957275, "learning_rate": 3.8540170390740097e-07, "loss": 0.1519, "step": 992 }, { "epoch": 0.9211502782931354, "grad_norm": 4.9389848709106445, "learning_rate": 3.7654733565969826e-07, "loss": 0.1874, "step": 993 }, { "epoch": 0.922077922077922, "grad_norm": 3.25769305229187, "learning_rate": 3.67793912876705e-07, "loss": 0.1191, "step": 994 }, { "epoch": 0.9230055658627088, "grad_norm": 3.4334826469421387, "learning_rate": 3.591415273776855e-07, "loss": 0.1012, "step": 995 }, { "epoch": 0.9239332096474954, "grad_norm": 3.1981468200683594, "learning_rate": 3.5059026992206645e-07, "loss": 0.0812, "step": 996 }, { "epoch": 0.924860853432282, "grad_norm": 5.118222236633301, "learning_rate": 3.421402302084953e-07, "loss": 0.1293, "step": 997 }, { "epoch": 0.9257884972170687, "grad_norm": 4.047184944152832, "learning_rate": 3.3379149687388866e-07, "loss": 0.1723, "step": 998 }, { "epoch": 0.9267161410018553, "grad_norm": 7.083133220672607, "learning_rate": 3.255441574925089e-07, "loss": 0.2061, "step": 999 }, { "epoch": 0.9276437847866419, "grad_norm": 2.8097355365753174, "learning_rate": 3.1739829857504235e-07, "loss": 0.086, "step": 1000 }, { "epoch": 0.9276437847866419, "eval_accuracy": 0.8580931263858093, "eval_f1": 0.6966824644549763, "eval_loss": 0.30505669116973877, "eval_precision": 0.8698224852071006, "eval_recall": 0.5810276679841897, "eval_runtime": 47.8654, "eval_samples_per_second": 5.766, "eval_steps_per_second": 0.188, "step": 1000 }, { "epoch": 0.9285714285714286, "grad_norm": 6.499805927276611, "learning_rate": 3.093540055676958e-07, "loss": 0.2015, "step": 1001 }, { "epoch": 0.9294990723562152, "grad_norm": 4.742324352264404, "learning_rate": 3.0141136285129825e-07, "loss": 0.1585, "step": 1002 }, { "epoch": 0.9304267161410018, "grad_norm": 4.395940780639648, "learning_rate": 2.935704537404083e-07, "loss": 0.1249, "step": 1003 }, { "epoch": 0.9313543599257885, "grad_norm": 3.036573886871338, "learning_rate": 2.8583136048245697e-07, "loss": 0.1331, "step": 1004 }, { "epoch": 0.9322820037105751, "grad_norm": 4.401485919952393, "learning_rate": 2.781941642568686e-07, "loss": 0.2138, "step": 1005 }, { "epoch": 0.9332096474953617, "grad_norm": 4.973133087158203, "learning_rate": 2.706589451742181e-07, "loss": 0.2253, "step": 1006 }, { "epoch": 0.9341372912801484, "grad_norm": 6.711733818054199, "learning_rate": 2.632257822753881e-07, "loss": 0.2465, "step": 1007 }, { "epoch": 0.935064935064935, "grad_norm": 3.2245848178863525, "learning_rate": 2.5589475353073987e-07, "loss": 0.1524, "step": 1008 }, { "epoch": 0.9359925788497218, "grad_norm": 3.8495306968688965, "learning_rate": 2.486659358392951e-07, "loss": 0.1646, "step": 1009 }, { "epoch": 0.9369202226345084, "grad_norm": 5.713381290435791, "learning_rate": 2.4153940502793185e-07, "loss": 0.2161, "step": 1010 }, { "epoch": 0.937847866419295, "grad_norm": 3.7789957523345947, "learning_rate": 2.3451523585058756e-07, "loss": 0.1509, "step": 1011 }, { "epoch": 0.9387755102040817, "grad_norm": 4.0073652267456055, "learning_rate": 2.2759350198746978e-07, "loss": 0.1402, "step": 1012 }, { "epoch": 0.9397031539888683, "grad_norm": 3.3210415840148926, "learning_rate": 2.2077427604429435e-07, "loss": 0.1431, "step": 1013 }, { "epoch": 0.9406307977736549, "grad_norm": 5.2902750968933105, "learning_rate": 2.1405762955151178e-07, "loss": 0.1239, "step": 1014 }, { "epoch": 0.9415584415584416, "grad_norm": 5.931840419769287, "learning_rate": 2.0744363296356872e-07, "loss": 0.1965, "step": 1015 }, { "epoch": 0.9424860853432282, "grad_norm": 3.9065020084381104, "learning_rate": 2.009323556581566e-07, "loss": 0.1344, "step": 1016 }, { "epoch": 0.9434137291280148, "grad_norm": 3.194225311279297, "learning_rate": 1.9452386593549534e-07, "loss": 0.0979, "step": 1017 }, { "epoch": 0.9443413729128015, "grad_norm": 3.794304847717285, "learning_rate": 1.8821823101760949e-07, "loss": 0.2038, "step": 1018 }, { "epoch": 0.9452690166975881, "grad_norm": 3.638219118118286, "learning_rate": 1.8201551704762453e-07, "loss": 0.1254, "step": 1019 }, { "epoch": 0.9461966604823747, "grad_norm": 4.820856094360352, "learning_rate": 1.7591578908907724e-07, "loss": 0.1909, "step": 1020 }, { "epoch": 0.9461966604823747, "eval_accuracy": 0.8580931263858093, "eval_f1": 0.6966824644549763, "eval_loss": 0.3055438697338104, "eval_precision": 0.8698224852071006, "eval_recall": 0.5810276679841897, "eval_runtime": 47.7431, "eval_samples_per_second": 5.781, "eval_steps_per_second": 0.189, "step": 1020 }, { "epoch": 0.9471243042671614, "grad_norm": 3.5032293796539307, "learning_rate": 1.699191111252241e-07, "loss": 0.077, "step": 1021 }, { "epoch": 0.948051948051948, "grad_norm": 3.7761011123657227, "learning_rate": 1.6402554605838173e-07, "loss": 0.1564, "step": 1022 }, { "epoch": 0.9489795918367347, "grad_norm": 5.951882839202881, "learning_rate": 1.5823515570925763e-07, "loss": 0.2323, "step": 1023 }, { "epoch": 0.9499072356215214, "grad_norm": 4.935650825500488, "learning_rate": 1.5254800081630828e-07, "loss": 0.2172, "step": 1024 }, { "epoch": 0.950834879406308, "grad_norm": 5.279281139373779, "learning_rate": 1.469641410350964e-07, "loss": 0.1845, "step": 1025 }, { "epoch": 0.9517625231910947, "grad_norm": 5.034005641937256, "learning_rate": 1.4148363493766803e-07, "loss": 0.174, "step": 1026 }, { "epoch": 0.9526901669758813, "grad_norm": 3.964360237121582, "learning_rate": 1.361065400119399e-07, "loss": 0.0888, "step": 1027 }, { "epoch": 0.9536178107606679, "grad_norm": 4.862616062164307, "learning_rate": 1.30832912661093e-07, "loss": 0.2007, "step": 1028 }, { "epoch": 0.9545454545454546, "grad_norm": 4.140252590179443, "learning_rate": 1.2566280820298427e-07, "loss": 0.13, "step": 1029 }, { "epoch": 0.9554730983302412, "grad_norm": 5.299205780029297, "learning_rate": 1.2059628086956044e-07, "loss": 0.1795, "step": 1030 }, { "epoch": 0.9564007421150278, "grad_norm": 5.694372653961182, "learning_rate": 1.1563338380629618e-07, "loss": 0.2278, "step": 1031 }, { "epoch": 0.9573283858998145, "grad_norm": 4.10621452331543, "learning_rate": 1.1077416907163573e-07, "loss": 0.1832, "step": 1032 }, { "epoch": 0.9582560296846011, "grad_norm": 3.0105836391448975, "learning_rate": 1.0601868763643997e-07, "loss": 0.1126, "step": 1033 }, { "epoch": 0.9591836734693877, "grad_norm": 5.412391185760498, "learning_rate": 1.0136698938346012e-07, "loss": 0.1996, "step": 1034 }, { "epoch": 0.9601113172541744, "grad_norm": 3.507596492767334, "learning_rate": 9.68191231068083e-08, "loss": 0.1647, "step": 1035 }, { "epoch": 0.961038961038961, "grad_norm": 4.733442783355713, "learning_rate": 9.237513651145224e-08, "loss": 0.102, "step": 1036 }, { "epoch": 0.9619666048237476, "grad_norm": 6.855641841888428, "learning_rate": 8.80350762127058e-08, "loss": 0.1197, "step": 1037 }, { "epoch": 0.9628942486085343, "grad_norm": 3.809262275695801, "learning_rate": 8.379898773574924e-08, "loss": 0.1287, "step": 1038 }, { "epoch": 0.963821892393321, "grad_norm": 3.4764761924743652, "learning_rate": 7.966691551514527e-08, "loss": 0.1106, "step": 1039 }, { "epoch": 0.9647495361781077, "grad_norm": 5.395627021789551, "learning_rate": 7.563890289437825e-08, "loss": 0.2017, "step": 1040 }, { "epoch": 0.9647495361781077, "eval_accuracy": 0.8580931263858093, "eval_f1": 0.6952380952380952, "eval_loss": 0.30581432580947876, "eval_precision": 0.874251497005988, "eval_recall": 0.5770750988142292, "eval_runtime": 48.746, "eval_samples_per_second": 5.662, "eval_steps_per_second": 0.185, "step": 1040 }, { "epoch": 0.9656771799628943, "grad_norm": 3.9932026863098145, "learning_rate": 7.171499212539124e-08, "loss": 0.1513, "step": 1041 }, { "epoch": 0.9666048237476809, "grad_norm": 5.8107075691223145, "learning_rate": 6.78952243681541e-08, "loss": 0.172, "step": 1042 }, { "epoch": 0.9675324675324676, "grad_norm": 3.7357017993927, "learning_rate": 6.417963969022389e-08, "loss": 0.1422, "step": 1043 }, { "epoch": 0.9684601113172542, "grad_norm": 3.854876756668091, "learning_rate": 6.056827706632185e-08, "loss": 0.1587, "step": 1044 }, { "epoch": 0.9693877551020408, "grad_norm": 6.006348133087158, "learning_rate": 5.7061174377937015e-08, "loss": 0.2244, "step": 1045 }, { "epoch": 0.9703153988868275, "grad_norm": 4.745636463165283, "learning_rate": 5.365836841291439e-08, "loss": 0.1803, "step": 1046 }, { "epoch": 0.9712430426716141, "grad_norm": 3.8510711193084717, "learning_rate": 5.035989486508075e-08, "loss": 0.1635, "step": 1047 }, { "epoch": 0.9721706864564007, "grad_norm": 5.504276752471924, "learning_rate": 4.716578833386054e-08, "loss": 0.1517, "step": 1048 }, { "epoch": 0.9730983302411874, "grad_norm": 3.3400299549102783, "learning_rate": 4.4076082323920576e-08, "loss": 0.1494, "step": 1049 }, { "epoch": 0.974025974025974, "grad_norm": 5.584471225738525, "learning_rate": 4.109080924481479e-08, "loss": 0.1781, "step": 1050 }, { "epoch": 0.9749536178107606, "grad_norm": 4.575666904449463, "learning_rate": 3.82100004106456e-08, "loss": 0.1298, "step": 1051 }, { "epoch": 0.9758812615955473, "grad_norm": 3.4456560611724854, "learning_rate": 3.543368603973529e-08, "loss": 0.1292, "step": 1052 }, { "epoch": 0.9768089053803339, "grad_norm": 2.841853618621826, "learning_rate": 3.2761895254306285e-08, "loss": 0.076, "step": 1053 }, { "epoch": 0.9777365491651205, "grad_norm": 4.662397384643555, "learning_rate": 3.019465608018024e-08, "loss": 0.2181, "step": 1054 }, { "epoch": 0.9786641929499073, "grad_norm": 3.259526014328003, "learning_rate": 2.773199544648164e-08, "loss": 0.108, "step": 1055 }, { "epoch": 0.9795918367346939, "grad_norm": 4.32330322265625, "learning_rate": 2.537393918535358e-08, "loss": 0.2068, "step": 1056 }, { "epoch": 0.9805194805194806, "grad_norm": 4.7618536949157715, "learning_rate": 2.312051203169352e-08, "loss": 0.1936, "step": 1057 }, { "epoch": 0.9814471243042672, "grad_norm": 4.779612064361572, "learning_rate": 2.0971737622883515e-08, "loss": 0.1007, "step": 1058 }, { "epoch": 0.9823747680890538, "grad_norm": 4.346301078796387, "learning_rate": 1.8927638498551502e-08, "loss": 0.1594, "step": 1059 }, { "epoch": 0.9833024118738405, "grad_norm": 4.017016410827637, "learning_rate": 1.698823610032929e-08, "loss": 0.1828, "step": 1060 }, { "epoch": 0.9833024118738405, "eval_accuracy": 0.8603104212860311, "eval_f1": 0.7028301886792453, "eval_loss": 0.306577205657959, "eval_precision": 0.8713450292397661, "eval_recall": 0.5889328063241107, "eval_runtime": 47.7001, "eval_samples_per_second": 5.786, "eval_steps_per_second": 0.189, "step": 1060 }, { "epoch": 0.9842300556586271, "grad_norm": 5.6250505447387695, "learning_rate": 1.5153550771630498e-08, "loss": 0.17, "step": 1061 }, { "epoch": 0.9851576994434137, "grad_norm": 5.426425933837891, "learning_rate": 1.3423601757436289e-08, "loss": 0.2122, "step": 1062 }, { "epoch": 0.9860853432282004, "grad_norm": 4.71135950088501, "learning_rate": 1.179840720409331e-08, "loss": 0.1715, "step": 1063 }, { "epoch": 0.987012987012987, "grad_norm": 4.350978851318359, "learning_rate": 1.0277984159122734e-08, "loss": 0.1704, "step": 1064 }, { "epoch": 0.9879406307977736, "grad_norm": 5.211360931396484, "learning_rate": 8.862348571043733e-09, "loss": 0.166, "step": 1065 }, { "epoch": 0.9888682745825603, "grad_norm": 4.015779495239258, "learning_rate": 7.551515289203615e-09, "loss": 0.1616, "step": 1066 }, { "epoch": 0.9897959183673469, "grad_norm": 4.356948375701904, "learning_rate": 6.345498063622391e-09, "loss": 0.1961, "step": 1067 }, { "epoch": 0.9907235621521335, "grad_norm": 6.508297920227051, "learning_rate": 5.2443095448506674e-09, "loss": 0.1935, "step": 1068 }, { "epoch": 0.9916512059369202, "grad_norm": 4.595229625701904, "learning_rate": 4.247961283835311e-09, "loss": 0.2116, "step": 1069 }, { "epoch": 0.9925788497217068, "grad_norm": 4.39501428604126, "learning_rate": 3.3564637317984318e-09, "loss": 0.1568, "step": 1070 }, { "epoch": 0.9935064935064936, "grad_norm": 4.391909599304199, "learning_rate": 2.5698262401263607e-09, "loss": 0.1553, "step": 1071 }, { "epoch": 0.9944341372912802, "grad_norm": 3.195699453353882, "learning_rate": 1.888057060274173e-09, "loss": 0.1469, "step": 1072 }, { "epoch": 0.9953617810760668, "grad_norm": 5.479938507080078, "learning_rate": 1.3111633436779792e-09, "loss": 0.1662, "step": 1073 }, { "epoch": 0.9962894248608535, "grad_norm": 4.181588172912598, "learning_rate": 8.391511416816489e-10, "loss": 0.1746, "step": 1074 }, { "epoch": 0.9972170686456401, "grad_norm": 4.144800662994385, "learning_rate": 4.720254054679796e-10, "loss": 0.1624, "step": 1075 }, { "epoch": 0.9981447124304267, "grad_norm": 3.857682228088379, "learning_rate": 2.0978998601206558e-10, "loss": 0.1293, "step": 1076 }, { "epoch": 0.9990723562152134, "grad_norm": 5.734769344329834, "learning_rate": 5.244763404133046e-11, "loss": 0.1897, "step": 1077 }, { "epoch": 1.0, "grad_norm": 3.7049508094787598, "learning_rate": 0.0, "loss": 0.1706, "step": 1078 } ], "logging_steps": 1, "max_steps": 1078, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.4499307937307034e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }