| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.999163179916318, | |
| "eval_steps": 20, | |
| "global_step": 796, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0, | |
| "eval_accuracy": 0.7339955849889624, | |
| "eval_f1": 0.2445141065830721, | |
| "eval_loss": 0.5994934439659119, | |
| "eval_precision": 0.6, | |
| "eval_recall": 0.15354330708661418, | |
| "eval_runtime": 52.2305, | |
| "eval_samples_per_second": 5.303, | |
| "eval_steps_per_second": 0.172, | |
| "step": 0 | |
| }, | |
| { | |
| "epoch": 0.0012552301255230125, | |
| "grad_norm": 1.9609311819076538, | |
| "learning_rate": 2.5000000000000004e-07, | |
| "loss": 0.6978, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.002510460251046025, | |
| "grad_norm": 1.9798370599746704, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 0.6725, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0037656903765690376, | |
| "grad_norm": 2.634610652923584, | |
| "learning_rate": 7.5e-07, | |
| "loss": 0.7798, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.00502092050209205, | |
| "grad_norm": 1.9435328245162964, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 0.6777, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.006276150627615063, | |
| "grad_norm": 2.0514488220214844, | |
| "learning_rate": 1.25e-06, | |
| "loss": 0.7849, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.007531380753138075, | |
| "grad_norm": 3.1893956661224365, | |
| "learning_rate": 1.5e-06, | |
| "loss": 0.7281, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.008786610878661089, | |
| "grad_norm": 2.2875595092773438, | |
| "learning_rate": 1.75e-06, | |
| "loss": 0.7267, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0100418410041841, | |
| "grad_norm": 1.7282941341400146, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 0.642, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.011297071129707114, | |
| "grad_norm": 1.9356372356414795, | |
| "learning_rate": 2.25e-06, | |
| "loss": 0.6923, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.012552301255230125, | |
| "grad_norm": 2.0085136890411377, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.6507, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.013807531380753139, | |
| "grad_norm": 2.1001739501953125, | |
| "learning_rate": 2.7500000000000004e-06, | |
| "loss": 0.6162, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.01506276150627615, | |
| "grad_norm": 2.053370714187622, | |
| "learning_rate": 3e-06, | |
| "loss": 0.674, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.016317991631799162, | |
| "grad_norm": 2.3665823936462402, | |
| "learning_rate": 3.2500000000000002e-06, | |
| "loss": 0.6931, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.017573221757322177, | |
| "grad_norm": 1.99113929271698, | |
| "learning_rate": 3.5e-06, | |
| "loss": 0.7218, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.01882845188284519, | |
| "grad_norm": 1.8170547485351562, | |
| "learning_rate": 3.7500000000000005e-06, | |
| "loss": 0.6488, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0200836820083682, | |
| "grad_norm": 1.8335210084915161, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.6467, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.021338912133891212, | |
| "grad_norm": 1.6069227457046509, | |
| "learning_rate": 4.25e-06, | |
| "loss": 0.6395, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.022594142259414227, | |
| "grad_norm": 1.7745310068130493, | |
| "learning_rate": 4.5e-06, | |
| "loss": 0.6905, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.02384937238493724, | |
| "grad_norm": 2.1716341972351074, | |
| "learning_rate": 4.75e-06, | |
| "loss": 0.653, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.02510460251046025, | |
| "grad_norm": 1.9034003019332886, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7266, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.02510460251046025, | |
| "eval_accuracy": 0.7384105960264901, | |
| "eval_f1": 0.26625386996904027, | |
| "eval_loss": 0.5908851027488708, | |
| "eval_precision": 0.6231884057971014, | |
| "eval_recall": 0.16929133858267717, | |
| "eval_runtime": 52.6006, | |
| "eval_samples_per_second": 5.266, | |
| "eval_steps_per_second": 0.171, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.026359832635983262, | |
| "grad_norm": 1.7850794792175293, | |
| "learning_rate": 5.2500000000000006e-06, | |
| "loss": 0.65, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.027615062761506277, | |
| "grad_norm": 1.765768051147461, | |
| "learning_rate": 5.500000000000001e-06, | |
| "loss": 0.6715, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.02887029288702929, | |
| "grad_norm": 1.8612842559814453, | |
| "learning_rate": 5.75e-06, | |
| "loss": 0.6614, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.0301255230125523, | |
| "grad_norm": 1.8554290533065796, | |
| "learning_rate": 6e-06, | |
| "loss": 0.641, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.03138075313807531, | |
| "grad_norm": 1.8867026567459106, | |
| "learning_rate": 6.25e-06, | |
| "loss": 0.6224, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.032635983263598324, | |
| "grad_norm": 1.8916778564453125, | |
| "learning_rate": 6.5000000000000004e-06, | |
| "loss": 0.6471, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.033891213389121336, | |
| "grad_norm": 1.7810847759246826, | |
| "learning_rate": 6.750000000000001e-06, | |
| "loss": 0.6197, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.035146443514644354, | |
| "grad_norm": 1.6231898069381714, | |
| "learning_rate": 7e-06, | |
| "loss": 0.6362, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.036401673640167366, | |
| "grad_norm": 1.7416272163391113, | |
| "learning_rate": 7.25e-06, | |
| "loss": 0.6425, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.03765690376569038, | |
| "grad_norm": 1.6757186651229858, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 0.5854, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03891213389121339, | |
| "grad_norm": 1.9119001626968384, | |
| "learning_rate": 7.75e-06, | |
| "loss": 0.5457, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.0401673640167364, | |
| "grad_norm": 1.576582908630371, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 0.5516, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.04142259414225941, | |
| "grad_norm": 1.5435791015625, | |
| "learning_rate": 8.25e-06, | |
| "loss": 0.5458, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.042677824267782424, | |
| "grad_norm": 1.8229247331619263, | |
| "learning_rate": 8.5e-06, | |
| "loss": 0.6167, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.043933054393305436, | |
| "grad_norm": 2.217472791671753, | |
| "learning_rate": 8.750000000000001e-06, | |
| "loss": 0.6588, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.045188284518828455, | |
| "grad_norm": 1.8096412420272827, | |
| "learning_rate": 9e-06, | |
| "loss": 0.6595, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.046443514644351466, | |
| "grad_norm": 1.8887217044830322, | |
| "learning_rate": 9.250000000000001e-06, | |
| "loss": 0.6004, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.04769874476987448, | |
| "grad_norm": 1.55510413646698, | |
| "learning_rate": 9.5e-06, | |
| "loss": 0.5685, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.04895397489539749, | |
| "grad_norm": 1.6017107963562012, | |
| "learning_rate": 9.75e-06, | |
| "loss": 0.497, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.0502092050209205, | |
| "grad_norm": 2.4420340061187744, | |
| "learning_rate": 1e-05, | |
| "loss": 0.674, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0502092050209205, | |
| "eval_accuracy": 0.7516556291390728, | |
| "eval_f1": 0.3553008595988539, | |
| "eval_loss": 0.5497225522994995, | |
| "eval_precision": 0.6526315789473685, | |
| "eval_recall": 0.2440944881889764, | |
| "eval_runtime": 53.1059, | |
| "eval_samples_per_second": 5.216, | |
| "eval_steps_per_second": 0.169, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.05146443514644351, | |
| "grad_norm": 1.6397217512130737, | |
| "learning_rate": 1.025e-05, | |
| "loss": 0.5409, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.052719665271966525, | |
| "grad_norm": 1.616377353668213, | |
| "learning_rate": 1.0500000000000001e-05, | |
| "loss": 0.5902, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.05397489539748954, | |
| "grad_norm": 1.5098403692245483, | |
| "learning_rate": 1.075e-05, | |
| "loss": 0.5468, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.055230125523012555, | |
| "grad_norm": 2.074469566345215, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 0.5969, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.056485355648535567, | |
| "grad_norm": 1.646159291267395, | |
| "learning_rate": 1.125e-05, | |
| "loss": 0.5629, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.05774058577405858, | |
| "grad_norm": 1.7806731462478638, | |
| "learning_rate": 1.15e-05, | |
| "loss": 0.565, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.05899581589958159, | |
| "grad_norm": 1.5080231428146362, | |
| "learning_rate": 1.1750000000000001e-05, | |
| "loss": 0.5349, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.0602510460251046, | |
| "grad_norm": 1.4364521503448486, | |
| "learning_rate": 1.2e-05, | |
| "loss": 0.5335, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.06150627615062761, | |
| "grad_norm": 1.3679174184799194, | |
| "learning_rate": 1.2250000000000001e-05, | |
| "loss": 0.5727, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.06276150627615062, | |
| "grad_norm": 2.1571309566497803, | |
| "learning_rate": 1.25e-05, | |
| "loss": 0.473, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.06401673640167364, | |
| "grad_norm": 1.4333269596099854, | |
| "learning_rate": 1.275e-05, | |
| "loss": 0.5388, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.06527196652719665, | |
| "grad_norm": 1.5196632146835327, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 0.5264, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.06652719665271967, | |
| "grad_norm": 1.491036057472229, | |
| "learning_rate": 1.325e-05, | |
| "loss": 0.4879, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.06778242677824267, | |
| "grad_norm": 2.134739875793457, | |
| "learning_rate": 1.3500000000000001e-05, | |
| "loss": 0.563, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.06903765690376569, | |
| "grad_norm": 1.4758329391479492, | |
| "learning_rate": 1.375e-05, | |
| "loss": 0.4958, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.07029288702928871, | |
| "grad_norm": 1.4116944074630737, | |
| "learning_rate": 1.4e-05, | |
| "loss": 0.5011, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.07154811715481171, | |
| "grad_norm": 2.1076178550720215, | |
| "learning_rate": 1.425e-05, | |
| "loss": 0.5847, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.07280334728033473, | |
| "grad_norm": 1.6353111267089844, | |
| "learning_rate": 1.45e-05, | |
| "loss": 0.573, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.07405857740585774, | |
| "grad_norm": 1.8148682117462158, | |
| "learning_rate": 1.4750000000000003e-05, | |
| "loss": 0.4973, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.07531380753138076, | |
| "grad_norm": 1.46212637424469, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.5187, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.07531380753138076, | |
| "eval_accuracy": 0.7759381898454746, | |
| "eval_f1": 0.5333333333333333, | |
| "eval_loss": 0.4896416962146759, | |
| "eval_precision": 0.6408839779005525, | |
| "eval_recall": 0.4566929133858268, | |
| "eval_runtime": 52.5243, | |
| "eval_samples_per_second": 5.274, | |
| "eval_steps_per_second": 0.171, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.07656903765690376, | |
| "grad_norm": 1.7205400466918945, | |
| "learning_rate": 1.525e-05, | |
| "loss": 0.4742, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.07782426778242678, | |
| "grad_norm": 1.6081304550170898, | |
| "learning_rate": 1.55e-05, | |
| "loss": 0.4835, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.0790794979079498, | |
| "grad_norm": 1.903638482093811, | |
| "learning_rate": 1.575e-05, | |
| "loss": 0.525, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.0803347280334728, | |
| "grad_norm": 1.9859038591384888, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 0.5183, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.08158995815899582, | |
| "grad_norm": 1.567049264907837, | |
| "learning_rate": 1.6250000000000002e-05, | |
| "loss": 0.4662, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.08284518828451883, | |
| "grad_norm": 1.9365341663360596, | |
| "learning_rate": 1.65e-05, | |
| "loss": 0.5389, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.08410041841004184, | |
| "grad_norm": 2.623508930206299, | |
| "learning_rate": 1.675e-05, | |
| "loss": 0.4868, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.08535564853556485, | |
| "grad_norm": 2.0028135776519775, | |
| "learning_rate": 1.7e-05, | |
| "loss": 0.5124, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.08661087866108787, | |
| "grad_norm": 1.8613665103912354, | |
| "learning_rate": 1.7250000000000003e-05, | |
| "loss": 0.4848, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.08786610878661087, | |
| "grad_norm": 2.0450963973999023, | |
| "learning_rate": 1.7500000000000002e-05, | |
| "loss": 0.4746, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.08912133891213389, | |
| "grad_norm": 2.370429515838623, | |
| "learning_rate": 1.775e-05, | |
| "loss": 0.4939, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.09037656903765691, | |
| "grad_norm": 2.433497905731201, | |
| "learning_rate": 1.8e-05, | |
| "loss": 0.3848, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.09163179916317991, | |
| "grad_norm": 2.043933868408203, | |
| "learning_rate": 1.825e-05, | |
| "loss": 0.4812, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.09288702928870293, | |
| "grad_norm": 3.1938834190368652, | |
| "learning_rate": 1.8500000000000002e-05, | |
| "loss": 0.4042, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.09414225941422594, | |
| "grad_norm": 2.2702953815460205, | |
| "learning_rate": 1.8750000000000002e-05, | |
| "loss": 0.4561, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.09539748953974896, | |
| "grad_norm": 2.0371036529541016, | |
| "learning_rate": 1.9e-05, | |
| "loss": 0.4152, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.09665271966527196, | |
| "grad_norm": 2.1548779010772705, | |
| "learning_rate": 1.925e-05, | |
| "loss": 0.4596, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.09790794979079498, | |
| "grad_norm": 2.221036195755005, | |
| "learning_rate": 1.95e-05, | |
| "loss": 0.3985, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.099163179916318, | |
| "grad_norm": 2.6363329887390137, | |
| "learning_rate": 1.9750000000000002e-05, | |
| "loss": 0.4098, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.100418410041841, | |
| "grad_norm": 3.167816162109375, | |
| "learning_rate": 2e-05, | |
| "loss": 0.4811, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.100418410041841, | |
| "eval_accuracy": 0.7958057395143487, | |
| "eval_f1": 0.5605700712589073, | |
| "eval_loss": 0.4410252869129181, | |
| "eval_precision": 0.7065868263473054, | |
| "eval_recall": 0.4645669291338583, | |
| "eval_runtime": 52.3433, | |
| "eval_samples_per_second": 5.292, | |
| "eval_steps_per_second": 0.172, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.10167364016736402, | |
| "grad_norm": 2.885763168334961, | |
| "learning_rate": 1.9999903740631467e-05, | |
| "loss": 0.4426, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.10292887029288703, | |
| "grad_norm": 3.0578081607818604, | |
| "learning_rate": 1.9999614964379037e-05, | |
| "loss": 0.3756, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.10418410041841004, | |
| "grad_norm": 2.6213362216949463, | |
| "learning_rate": 1.9999133676802198e-05, | |
| "loss": 0.3389, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.10543933054393305, | |
| "grad_norm": 2.924165964126587, | |
| "learning_rate": 1.9998459887166635e-05, | |
| "loss": 0.4445, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.10669456066945607, | |
| "grad_norm": 3.2457644939422607, | |
| "learning_rate": 1.999759360844406e-05, | |
| "loss": 0.3953, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.10794979079497909, | |
| "grad_norm": 3.2740261554718018, | |
| "learning_rate": 1.9996534857311967e-05, | |
| "loss": 0.3838, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.10920502092050209, | |
| "grad_norm": 2.938350200653076, | |
| "learning_rate": 1.999528365415329e-05, | |
| "loss": 0.3549, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.11046025104602511, | |
| "grad_norm": 3.5133864879608154, | |
| "learning_rate": 1.9993840023056045e-05, | |
| "loss": 0.3628, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.11171548117154811, | |
| "grad_norm": 4.392985820770264, | |
| "learning_rate": 1.9992203991812823e-05, | |
| "loss": 0.3628, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.11297071129707113, | |
| "grad_norm": 4.105576515197754, | |
| "learning_rate": 1.9990375591920304e-05, | |
| "loss": 0.3535, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.11422594142259414, | |
| "grad_norm": 3.1926333904266357, | |
| "learning_rate": 1.9988354858578603e-05, | |
| "loss": 0.3401, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.11548117154811716, | |
| "grad_norm": 3.768364191055298, | |
| "learning_rate": 1.9986141830690626e-05, | |
| "loss": 0.3704, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.11673640167364016, | |
| "grad_norm": 7.29379940032959, | |
| "learning_rate": 1.9983736550861306e-05, | |
| "loss": 0.4349, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.11799163179916318, | |
| "grad_norm": 2.7373130321502686, | |
| "learning_rate": 1.9981139065396786e-05, | |
| "loss": 0.3264, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.1192468619246862, | |
| "grad_norm": 6.39415979385376, | |
| "learning_rate": 1.9978349424303532e-05, | |
| "loss": 0.3619, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.1205020920502092, | |
| "grad_norm": 3.354396343231201, | |
| "learning_rate": 1.9975367681287358e-05, | |
| "loss": 0.338, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.12175732217573222, | |
| "grad_norm": 4.022518634796143, | |
| "learning_rate": 1.99721938937524e-05, | |
| "loss": 0.3851, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.12301255230125523, | |
| "grad_norm": 3.3857271671295166, | |
| "learning_rate": 1.9968828122800022e-05, | |
| "loss": 0.352, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.12426778242677824, | |
| "grad_norm": 3.325467824935913, | |
| "learning_rate": 1.9965270433227623e-05, | |
| "loss": 0.3588, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.12552301255230125, | |
| "grad_norm": 3.3111820220947266, | |
| "learning_rate": 1.9961520893527385e-05, | |
| "loss": 0.2811, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.12552301255230125, | |
| "eval_accuracy": 0.8101545253863135, | |
| "eval_f1": 0.5825242718446602, | |
| "eval_loss": 0.4248526096343994, | |
| "eval_precision": 0.759493670886076, | |
| "eval_recall": 0.47244094488188976, | |
| "eval_runtime": 53.343, | |
| "eval_samples_per_second": 5.193, | |
| "eval_steps_per_second": 0.169, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.12677824267782425, | |
| "grad_norm": 3.9941301345825195, | |
| "learning_rate": 1.9957579575884978e-05, | |
| "loss": 0.3364, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.1280334728033473, | |
| "grad_norm": 3.4411442279815674, | |
| "learning_rate": 1.995344655617815e-05, | |
| "loss": 0.3481, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.1292887029288703, | |
| "grad_norm": 3.1757569313049316, | |
| "learning_rate": 1.9949121913975275e-05, | |
| "loss": 0.3447, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.1305439330543933, | |
| "grad_norm": 5.263054847717285, | |
| "learning_rate": 1.994460573253382e-05, | |
| "loss": 0.3705, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.13179916317991633, | |
| "grad_norm": 4.291073322296143, | |
| "learning_rate": 1.9939898098798736e-05, | |
| "loss": 0.349, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.13305439330543933, | |
| "grad_norm": 4.688785076141357, | |
| "learning_rate": 1.9934999103400797e-05, | |
| "loss": 0.2573, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.13430962343096234, | |
| "grad_norm": 3.483659505844116, | |
| "learning_rate": 1.992990884065484e-05, | |
| "loss": 0.2812, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.13556485355648534, | |
| "grad_norm": 5.222522258758545, | |
| "learning_rate": 1.9924627408557963e-05, | |
| "loss": 0.3208, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.13682008368200838, | |
| "grad_norm": 3.1946051120758057, | |
| "learning_rate": 1.991915490878763e-05, | |
| "loss": 0.4041, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.13807531380753138, | |
| "grad_norm": 2.834019899368286, | |
| "learning_rate": 1.9913491446699715e-05, | |
| "loss": 0.2989, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.13933054393305438, | |
| "grad_norm": 4.4058380126953125, | |
| "learning_rate": 1.9907637131326475e-05, | |
| "loss": 0.3247, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.14058577405857742, | |
| "grad_norm": 4.437101364135742, | |
| "learning_rate": 1.9901592075374447e-05, | |
| "loss": 0.3487, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.14184100418410042, | |
| "grad_norm": 3.1267802715301514, | |
| "learning_rate": 1.989535639522229e-05, | |
| "loss": 0.2741, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.14309623430962343, | |
| "grad_norm": 3.8325576782226562, | |
| "learning_rate": 1.988893021091853e-05, | |
| "loss": 0.385, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.14435146443514643, | |
| "grad_norm": 4.569618225097656, | |
| "learning_rate": 1.9882313646179247e-05, | |
| "loss": 0.3595, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.14560669456066946, | |
| "grad_norm": 5.401278972625732, | |
| "learning_rate": 1.9875506828385723e-05, | |
| "loss": 0.2875, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.14686192468619247, | |
| "grad_norm": 3.8402180671691895, | |
| "learning_rate": 1.9868509888581945e-05, | |
| "loss": 0.3079, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.14811715481171547, | |
| "grad_norm": 3.5551564693450928, | |
| "learning_rate": 1.986132296147212e-05, | |
| "loss": 0.3157, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.1493723849372385, | |
| "grad_norm": 9.136929512023926, | |
| "learning_rate": 1.9853946185418056e-05, | |
| "loss": 0.4209, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.1506276150627615, | |
| "grad_norm": 3.290203332901001, | |
| "learning_rate": 1.9846379702436518e-05, | |
| "loss": 0.2959, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.1506276150627615, | |
| "eval_accuracy": 0.8211920529801324, | |
| "eval_f1": 0.6197183098591549, | |
| "eval_loss": 0.37509337067604065, | |
| "eval_precision": 0.7674418604651163, | |
| "eval_recall": 0.5196850393700787, | |
| "eval_runtime": 52.5731, | |
| "eval_samples_per_second": 5.269, | |
| "eval_steps_per_second": 0.171, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.15188284518828452, | |
| "grad_norm": 4.515352725982666, | |
| "learning_rate": 1.983862365819648e-05, | |
| "loss": 0.3283, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.15313807531380752, | |
| "grad_norm": 3.97063946723938, | |
| "learning_rate": 1.9830678202016324e-05, | |
| "loss": 0.3505, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.15439330543933055, | |
| "grad_norm": 4.553818225860596, | |
| "learning_rate": 1.982254348686097e-05, | |
| "loss": 0.313, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.15564853556485356, | |
| "grad_norm": 3.5846359729766846, | |
| "learning_rate": 1.981421966933893e-05, | |
| "loss": 0.35, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.15690376569037656, | |
| "grad_norm": 5.479614734649658, | |
| "learning_rate": 1.9805706909699283e-05, | |
| "loss": 0.3134, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.1581589958158996, | |
| "grad_norm": 3.6926157474517822, | |
| "learning_rate": 1.9797005371828603e-05, | |
| "loss": 0.3659, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.1594142259414226, | |
| "grad_norm": 4.4174957275390625, | |
| "learning_rate": 1.97881152232478e-05, | |
| "loss": 0.3069, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.1606694560669456, | |
| "grad_norm": 2.855861186981201, | |
| "learning_rate": 1.9779036635108892e-05, | |
| "loss": 0.2748, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.1619246861924686, | |
| "grad_norm": 3.4113943576812744, | |
| "learning_rate": 1.976976978219171e-05, | |
| "loss": 0.2942, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.16317991631799164, | |
| "grad_norm": 2.8706114292144775, | |
| "learning_rate": 1.9760314842900537e-05, | |
| "loss": 0.26, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.16443514644351465, | |
| "grad_norm": 3.3289883136749268, | |
| "learning_rate": 1.975067199926067e-05, | |
| "loss": 0.2942, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.16569037656903765, | |
| "grad_norm": 3.1963343620300293, | |
| "learning_rate": 1.9740841436914917e-05, | |
| "loss": 0.3404, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.16694560669456068, | |
| "grad_norm": 4.106410026550293, | |
| "learning_rate": 1.9730823345120024e-05, | |
| "loss": 0.3645, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.1682008368200837, | |
| "grad_norm": 3.587475299835205, | |
| "learning_rate": 1.9720617916743022e-05, | |
| "loss": 0.2905, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.1694560669456067, | |
| "grad_norm": 4.643335819244385, | |
| "learning_rate": 1.971022534825754e-05, | |
| "loss": 0.3199, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.1707112970711297, | |
| "grad_norm": 3.8745625019073486, | |
| "learning_rate": 1.9699645839739987e-05, | |
| "loss": 0.3276, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.17196652719665273, | |
| "grad_norm": 4.443915367126465, | |
| "learning_rate": 1.9688879594865726e-05, | |
| "loss": 0.1989, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.17322175732217573, | |
| "grad_norm": 3.165154218673706, | |
| "learning_rate": 1.9677926820905143e-05, | |
| "loss": 0.2877, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.17447698744769874, | |
| "grad_norm": 3.396127462387085, | |
| "learning_rate": 1.9666787728719664e-05, | |
| "loss": 0.2869, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.17573221757322174, | |
| "grad_norm": 4.032714366912842, | |
| "learning_rate": 1.9655462532757677e-05, | |
| "loss": 0.336, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.17573221757322174, | |
| "eval_accuracy": 0.8278145695364238, | |
| "eval_f1": 0.6060606060606061, | |
| "eval_loss": 0.3764040172100067, | |
| "eval_precision": 0.8450704225352113, | |
| "eval_recall": 0.47244094488188976, | |
| "eval_runtime": 52.6872, | |
| "eval_samples_per_second": 5.257, | |
| "eval_steps_per_second": 0.171, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.17698744769874478, | |
| "grad_norm": 2.6727371215820312, | |
| "learning_rate": 1.9643951451050428e-05, | |
| "loss": 0.2636, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.17824267782426778, | |
| "grad_norm": 3.8816864490509033, | |
| "learning_rate": 1.9632254705207813e-05, | |
| "loss": 0.3208, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.17949790794979079, | |
| "grad_norm": 3.4616892337799072, | |
| "learning_rate": 1.9620372520414098e-05, | |
| "loss": 0.3218, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.18075313807531382, | |
| "grad_norm": 4.056252479553223, | |
| "learning_rate": 1.9608305125423608e-05, | |
| "loss": 0.2844, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.18200836820083682, | |
| "grad_norm": 5.902234077453613, | |
| "learning_rate": 1.9596052752556308e-05, | |
| "loss": 0.2497, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.18326359832635983, | |
| "grad_norm": 4.0488996505737305, | |
| "learning_rate": 1.958361563769333e-05, | |
| "loss": 0.2764, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.18451882845188283, | |
| "grad_norm": 4.619633197784424, | |
| "learning_rate": 1.957099402027244e-05, | |
| "loss": 0.3775, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.18577405857740587, | |
| "grad_norm": 4.491790294647217, | |
| "learning_rate": 1.9558188143283425e-05, | |
| "loss": 0.4185, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.18702928870292887, | |
| "grad_norm": 9.393437385559082, | |
| "learning_rate": 1.954519825326341e-05, | |
| "loss": 0.292, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.18828451882845187, | |
| "grad_norm": 9.774816513061523, | |
| "learning_rate": 1.9532024600292115e-05, | |
| "loss": 0.341, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.1895397489539749, | |
| "grad_norm": 9.051419258117676, | |
| "learning_rate": 1.9518667437987045e-05, | |
| "loss": 0.3125, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.1907949790794979, | |
| "grad_norm": 4.726169586181641, | |
| "learning_rate": 1.9505127023498603e-05, | |
| "loss": 0.3283, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.19205020920502092, | |
| "grad_norm": 3.818352222442627, | |
| "learning_rate": 1.9491403617505134e-05, | |
| "loss": 0.2696, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.19330543933054392, | |
| "grad_norm": 4.901086330413818, | |
| "learning_rate": 1.9477497484207922e-05, | |
| "loss": 0.2927, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.19456066945606695, | |
| "grad_norm": 2.7958414554595947, | |
| "learning_rate": 1.9463408891326088e-05, | |
| "loss": 0.2544, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.19581589958158996, | |
| "grad_norm": 3.541666030883789, | |
| "learning_rate": 1.9449138110091444e-05, | |
| "loss": 0.2723, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.19707112970711296, | |
| "grad_norm": 4.369930744171143, | |
| "learning_rate": 1.9434685415243267e-05, | |
| "loss": 0.3121, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.198326359832636, | |
| "grad_norm": 4.061751842498779, | |
| "learning_rate": 1.9420051085023006e-05, | |
| "loss": 0.3238, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.199581589958159, | |
| "grad_norm": 5.1077446937561035, | |
| "learning_rate": 1.940523540116895e-05, | |
| "loss": 0.2935, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.200836820083682, | |
| "grad_norm": 3.7316763401031494, | |
| "learning_rate": 1.9390238648910765e-05, | |
| "loss": 0.3239, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.200836820083682, | |
| "eval_accuracy": 0.8200883002207505, | |
| "eval_f1": 0.5788113695090439, | |
| "eval_loss": 0.3607686161994934, | |
| "eval_precision": 0.8421052631578947, | |
| "eval_recall": 0.4409448818897638, | |
| "eval_runtime": 53.735, | |
| "eval_samples_per_second": 5.155, | |
| "eval_steps_per_second": 0.167, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.202092050209205, | |
| "grad_norm": 4.719494819641113, | |
| "learning_rate": 1.9375061116964032e-05, | |
| "loss": 0.3164, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.20334728033472804, | |
| "grad_norm": 3.245194435119629, | |
| "learning_rate": 1.935970309752469e-05, | |
| "loss": 0.2923, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.20460251046025105, | |
| "grad_norm": 4.244296550750732, | |
| "learning_rate": 1.9344164886263375e-05, | |
| "loss": 0.2891, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.20585774058577405, | |
| "grad_norm": 5.457589149475098, | |
| "learning_rate": 1.932844678231977e-05, | |
| "loss": 0.3057, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.20711297071129708, | |
| "grad_norm": 4.439499378204346, | |
| "learning_rate": 1.9312549088296838e-05, | |
| "loss": 0.2107, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.2083682008368201, | |
| "grad_norm": 5.0200653076171875, | |
| "learning_rate": 1.929647211025497e-05, | |
| "loss": 0.2859, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.2096234309623431, | |
| "grad_norm": 3.7708117961883545, | |
| "learning_rate": 1.9280216157706113e-05, | |
| "loss": 0.2816, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.2108786610878661, | |
| "grad_norm": 3.947610855102539, | |
| "learning_rate": 1.9263781543607817e-05, | |
| "loss": 0.2431, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.21213389121338913, | |
| "grad_norm": 3.28195858001709, | |
| "learning_rate": 1.9247168584357195e-05, | |
| "loss": 0.296, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.21338912133891214, | |
| "grad_norm": 3.6983871459960938, | |
| "learning_rate": 1.923037759978484e-05, | |
| "loss": 0.3003, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.21464435146443514, | |
| "grad_norm": 4.456281661987305, | |
| "learning_rate": 1.921340891314867e-05, | |
| "loss": 0.2493, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.21589958158995817, | |
| "grad_norm": 3.2370941638946533, | |
| "learning_rate": 1.9196262851127695e-05, | |
| "loss": 0.2353, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.21715481171548118, | |
| "grad_norm": 2.977496862411499, | |
| "learning_rate": 1.9178939743815735e-05, | |
| "loss": 0.3062, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.21841004184100418, | |
| "grad_norm": 5.293909072875977, | |
| "learning_rate": 1.9161439924715063e-05, | |
| "loss": 0.2646, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.2196652719665272, | |
| "grad_norm": 3.4083428382873535, | |
| "learning_rate": 1.9143763730729987e-05, | |
| "loss": 0.2305, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.22092050209205022, | |
| "grad_norm": 2.7759830951690674, | |
| "learning_rate": 1.9125911502160365e-05, | |
| "loss": 0.2554, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.22217573221757322, | |
| "grad_norm": 3.9626009464263916, | |
| "learning_rate": 1.9107883582695043e-05, | |
| "loss": 0.2789, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.22343096234309623, | |
| "grad_norm": 3.340153932571411, | |
| "learning_rate": 1.9089680319405252e-05, | |
| "loss": 0.2874, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.22468619246861923, | |
| "grad_norm": 3.277308702468872, | |
| "learning_rate": 1.9071302062737915e-05, | |
| "loss": 0.1978, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.22594142259414227, | |
| "grad_norm": 5.420035362243652, | |
| "learning_rate": 1.905274916650891e-05, | |
| "loss": 0.2767, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.22594142259414227, | |
| "eval_accuracy": 0.8543046357615894, | |
| "eval_f1": 0.7066666666666667, | |
| "eval_loss": 0.3361983299255371, | |
| "eval_precision": 0.8112244897959183, | |
| "eval_recall": 0.6259842519685039, | |
| "eval_runtime": 51.9639, | |
| "eval_samples_per_second": 5.331, | |
| "eval_steps_per_second": 0.173, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.22719665271966527, | |
| "grad_norm": 6.328350067138672, | |
| "learning_rate": 1.903402198789625e-05, | |
| "loss": 0.3489, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.22845188284518828, | |
| "grad_norm": 3.141185760498047, | |
| "learning_rate": 1.9015120887433215e-05, | |
| "loss": 0.3043, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.2297071129707113, | |
| "grad_norm": 3.633781909942627, | |
| "learning_rate": 1.8996046229001407e-05, | |
| "loss": 0.3081, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.2309623430962343, | |
| "grad_norm": 3.6458773612976074, | |
| "learning_rate": 1.897679837982373e-05, | |
| "loss": 0.2259, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.23221757322175732, | |
| "grad_norm": 7.069568634033203, | |
| "learning_rate": 1.895737771045736e-05, | |
| "loss": 0.347, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.23347280334728032, | |
| "grad_norm": 5.888752460479736, | |
| "learning_rate": 1.8937784594786562e-05, | |
| "loss": 0.2361, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.23472803347280335, | |
| "grad_norm": 3.553389072418213, | |
| "learning_rate": 1.8918019410015527e-05, | |
| "loss": 0.2504, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.23598326359832636, | |
| "grad_norm": 3.6231913566589355, | |
| "learning_rate": 1.8898082536661097e-05, | |
| "loss": 0.2558, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.23723849372384936, | |
| "grad_norm": 3.3635237216949463, | |
| "learning_rate": 1.887797435854543e-05, | |
| "loss": 0.2867, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.2384937238493724, | |
| "grad_norm": 4.0168538093566895, | |
| "learning_rate": 1.885769526278865e-05, | |
| "loss": 0.3195, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2397489539748954, | |
| "grad_norm": 4.260074615478516, | |
| "learning_rate": 1.8837245639801332e-05, | |
| "loss": 0.2861, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.2410041841004184, | |
| "grad_norm": 3.299710988998413, | |
| "learning_rate": 1.8816625883277044e-05, | |
| "loss": 0.2454, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.2422594142259414, | |
| "grad_norm": 4.68196439743042, | |
| "learning_rate": 1.8795836390184727e-05, | |
| "loss": 0.2976, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.24351464435146444, | |
| "grad_norm": 4.414516925811768, | |
| "learning_rate": 1.8774877560761082e-05, | |
| "loss": 0.2814, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.24476987447698745, | |
| "grad_norm": 4.108029365539551, | |
| "learning_rate": 1.8753749798502845e-05, | |
| "loss": 0.2478, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.24602510460251045, | |
| "grad_norm": 3.553065061569214, | |
| "learning_rate": 1.8732453510159025e-05, | |
| "loss": 0.2221, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.24728033472803349, | |
| "grad_norm": 3.1897339820861816, | |
| "learning_rate": 1.871098910572308e-05, | |
| "loss": 0.3001, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.2485355648535565, | |
| "grad_norm": 4.416936874389648, | |
| "learning_rate": 1.8689356998425007e-05, | |
| "loss": 0.3109, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.2497907949790795, | |
| "grad_norm": 3.162482976913452, | |
| "learning_rate": 1.8667557604723404e-05, | |
| "loss": 0.3104, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.2510460251046025, | |
| "grad_norm": 3.864384651184082, | |
| "learning_rate": 1.864559134429745e-05, | |
| "loss": 0.276, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.2510460251046025, | |
| "eval_accuracy": 0.8388520971302428, | |
| "eval_f1": 0.6439024390243903, | |
| "eval_loss": 0.3405630886554718, | |
| "eval_precision": 0.8461538461538461, | |
| "eval_recall": 0.5196850393700787, | |
| "eval_runtime": 52.2841, | |
| "eval_samples_per_second": 5.298, | |
| "eval_steps_per_second": 0.172, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.25230125523012553, | |
| "grad_norm": 3.1597511768341064, | |
| "learning_rate": 1.8623458640038817e-05, | |
| "loss": 0.2417, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.2535564853556485, | |
| "grad_norm": 4.373691558837891, | |
| "learning_rate": 1.8601159918043533e-05, | |
| "loss": 0.2408, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.25481171548117154, | |
| "grad_norm": 3.5026726722717285, | |
| "learning_rate": 1.857869560760377e-05, | |
| "loss": 0.281, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.2560669456066946, | |
| "grad_norm": 4.196898460388184, | |
| "learning_rate": 1.85560661411996e-05, | |
| "loss": 0.2201, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.25732217573221755, | |
| "grad_norm": 3.8971402645111084, | |
| "learning_rate": 1.8533271954490655e-05, | |
| "loss": 0.2692, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.2585774058577406, | |
| "grad_norm": 6.247049808502197, | |
| "learning_rate": 1.8510313486307734e-05, | |
| "loss": 0.2653, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.2598326359832636, | |
| "grad_norm": 5.326446056365967, | |
| "learning_rate": 1.848719117864437e-05, | |
| "loss": 0.2857, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.2610878661087866, | |
| "grad_norm": 3.6153714656829834, | |
| "learning_rate": 1.846390547664831e-05, | |
| "loss": 0.262, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.2623430962343096, | |
| "grad_norm": 6.952093124389648, | |
| "learning_rate": 1.8440456828612946e-05, | |
| "loss": 0.2807, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.26359832635983266, | |
| "grad_norm": 3.8363044261932373, | |
| "learning_rate": 1.841684568596869e-05, | |
| "loss": 0.2604, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.26485355648535563, | |
| "grad_norm": 3.643761396408081, | |
| "learning_rate": 1.8393072503274277e-05, | |
| "loss": 0.2796, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.26610878661087867, | |
| "grad_norm": 3.259951114654541, | |
| "learning_rate": 1.836913773820802e-05, | |
| "loss": 0.2724, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.2673640167364017, | |
| "grad_norm": 4.189282417297363, | |
| "learning_rate": 1.834504185155899e-05, | |
| "loss": 0.2455, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.2686192468619247, | |
| "grad_norm": 4.426260948181152, | |
| "learning_rate": 1.832078530721816e-05, | |
| "loss": 0.2975, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.2698744769874477, | |
| "grad_norm": 4.503783226013184, | |
| "learning_rate": 1.829636857216945e-05, | |
| "loss": 0.2852, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.2711297071129707, | |
| "grad_norm": 4.618401527404785, | |
| "learning_rate": 1.8271792116480767e-05, | |
| "loss": 0.3006, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.2723849372384937, | |
| "grad_norm": 5.484090805053711, | |
| "learning_rate": 1.8247056413294927e-05, | |
| "loss": 0.3397, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.27364016736401675, | |
| "grad_norm": 4.215097427368164, | |
| "learning_rate": 1.8222161938820564e-05, | |
| "loss": 0.312, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.27489539748953973, | |
| "grad_norm": 2.8045787811279297, | |
| "learning_rate": 1.8197109172322958e-05, | |
| "loss": 0.2896, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.27615062761506276, | |
| "grad_norm": 3.3742685317993164, | |
| "learning_rate": 1.8171898596114804e-05, | |
| "loss": 0.2715, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.27615062761506276, | |
| "eval_accuracy": 0.8410596026490066, | |
| "eval_f1": 0.6587677725118484, | |
| "eval_loss": 0.3223263919353485, | |
| "eval_precision": 0.8273809523809523, | |
| "eval_recall": 0.547244094488189, | |
| "eval_runtime": 52.2317, | |
| "eval_samples_per_second": 5.303, | |
| "eval_steps_per_second": 0.172, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.2774058577405858, | |
| "grad_norm": 2.477954149246216, | |
| "learning_rate": 1.8146530695546934e-05, | |
| "loss": 0.2171, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.27866108786610877, | |
| "grad_norm": 3.73885440826416, | |
| "learning_rate": 1.8121005958998968e-05, | |
| "loss": 0.3282, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.2799163179916318, | |
| "grad_norm": 3.0945334434509277, | |
| "learning_rate": 1.8095324877869902e-05, | |
| "loss": 0.2823, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.28117154811715483, | |
| "grad_norm": 3.98866868019104, | |
| "learning_rate": 1.8069487946568675e-05, | |
| "loss": 0.3008, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.2824267782426778, | |
| "grad_norm": 3.4288768768310547, | |
| "learning_rate": 1.804349566250462e-05, | |
| "loss": 0.2644, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.28368200836820084, | |
| "grad_norm": 3.3643836975097656, | |
| "learning_rate": 1.801734852607791e-05, | |
| "loss": 0.2543, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.2849372384937239, | |
| "grad_norm": 5.725021839141846, | |
| "learning_rate": 1.799104704066991e-05, | |
| "loss": 0.2827, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.28619246861924685, | |
| "grad_norm": 4.484889507293701, | |
| "learning_rate": 1.79645917126335e-05, | |
| "loss": 0.3096, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.2874476987447699, | |
| "grad_norm": 5.622531414031982, | |
| "learning_rate": 1.7937983051283312e-05, | |
| "loss": 0.3283, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.28870292887029286, | |
| "grad_norm": 4.898491382598877, | |
| "learning_rate": 1.7911221568885935e-05, | |
| "loss": 0.2316, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2899581589958159, | |
| "grad_norm": 4.367154121398926, | |
| "learning_rate": 1.7884307780650047e-05, | |
| "loss": 0.2739, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.29121338912133893, | |
| "grad_norm": 4.016841888427734, | |
| "learning_rate": 1.7857242204716497e-05, | |
| "loss": 0.2375, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.2924686192468619, | |
| "grad_norm": 3.615976333618164, | |
| "learning_rate": 1.783002536214834e-05, | |
| "loss": 0.2644, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.29372384937238494, | |
| "grad_norm": 5.212274074554443, | |
| "learning_rate": 1.780265777692079e-05, | |
| "loss": 0.3412, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.29497907949790797, | |
| "grad_norm": 3.3848087787628174, | |
| "learning_rate": 1.7775139975911143e-05, | |
| "loss": 0.2489, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.29623430962343095, | |
| "grad_norm": 5.973453998565674, | |
| "learning_rate": 1.7747472488888622e-05, | |
| "loss": 0.2657, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.297489539748954, | |
| "grad_norm": 4.158175468444824, | |
| "learning_rate": 1.77196558485042e-05, | |
| "loss": 0.2951, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.298744769874477, | |
| "grad_norm": 3.3108043670654297, | |
| "learning_rate": 1.7691690590280325e-05, | |
| "loss": 0.26, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 4.062819480895996, | |
| "learning_rate": 1.7663577252600612e-05, | |
| "loss": 0.2535, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.301255230125523, | |
| "grad_norm": 4.0478339195251465, | |
| "learning_rate": 1.763531637669949e-05, | |
| "loss": 0.2737, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.301255230125523, | |
| "eval_accuracy": 0.8520971302428256, | |
| "eval_f1": 0.6995515695067265, | |
| "eval_loss": 0.3201642632484436, | |
| "eval_precision": 0.8125, | |
| "eval_recall": 0.6141732283464567, | |
| "eval_runtime": 52.8485, | |
| "eval_samples_per_second": 5.241, | |
| "eval_steps_per_second": 0.17, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.302510460251046, | |
| "grad_norm": 5.782260894775391, | |
| "learning_rate": 1.760690850665177e-05, | |
| "loss": 0.2356, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.30376569037656903, | |
| "grad_norm": 4.108422756195068, | |
| "learning_rate": 1.7578354189362183e-05, | |
| "loss": 0.2658, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.30502092050209206, | |
| "grad_norm": 2.872807264328003, | |
| "learning_rate": 1.7549653974554835e-05, | |
| "loss": 0.3048, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.30627615062761504, | |
| "grad_norm": 3.7681846618652344, | |
| "learning_rate": 1.752080841476264e-05, | |
| "loss": 0.2832, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.3075313807531381, | |
| "grad_norm": 6.7302069664001465, | |
| "learning_rate": 1.7491818065316676e-05, | |
| "loss": 0.2518, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.3087866108786611, | |
| "grad_norm": 7.851168155670166, | |
| "learning_rate": 1.7462683484335477e-05, | |
| "loss": 0.2188, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.3100418410041841, | |
| "grad_norm": 5.26230001449585, | |
| "learning_rate": 1.7433405232714325e-05, | |
| "loss": 0.2898, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.3112970711297071, | |
| "grad_norm": 3.618230104446411, | |
| "learning_rate": 1.7403983874114422e-05, | |
| "loss": 0.2303, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.31255230125523015, | |
| "grad_norm": 3.8040518760681152, | |
| "learning_rate": 1.7374419974952045e-05, | |
| "loss": 0.3179, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.3138075313807531, | |
| "grad_norm": 3.1975717544555664, | |
| "learning_rate": 1.734471410438765e-05, | |
| "loss": 0.2503, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.31506276150627616, | |
| "grad_norm": 2.8378207683563232, | |
| "learning_rate": 1.731486683431491e-05, | |
| "loss": 0.2424, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.3163179916317992, | |
| "grad_norm": 5.816548824310303, | |
| "learning_rate": 1.728487873934969e-05, | |
| "loss": 0.2567, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.31757322175732217, | |
| "grad_norm": 3.5895259380340576, | |
| "learning_rate": 1.7254750396819008e-05, | |
| "loss": 0.2762, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.3188284518828452, | |
| "grad_norm": 3.293178081512451, | |
| "learning_rate": 1.7224482386749916e-05, | |
| "loss": 0.2801, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.3200836820083682, | |
| "grad_norm": 3.76770281791687, | |
| "learning_rate": 1.719407529185831e-05, | |
| "loss": 0.2545, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.3213389121338912, | |
| "grad_norm": 3.1176042556762695, | |
| "learning_rate": 1.7163529697537756e-05, | |
| "loss": 0.2608, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.32259414225941424, | |
| "grad_norm": 3.789315700531006, | |
| "learning_rate": 1.7132846191848167e-05, | |
| "loss": 0.2708, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.3238493723849372, | |
| "grad_norm": 5.817142963409424, | |
| "learning_rate": 1.7102025365504524e-05, | |
| "loss": 0.3254, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.32510460251046025, | |
| "grad_norm": 4.174067497253418, | |
| "learning_rate": 1.7071067811865477e-05, | |
| "loss": 0.2826, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.3263598326359833, | |
| "grad_norm": 4.383941173553467, | |
| "learning_rate": 1.7039974126921946e-05, | |
| "loss": 0.3245, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.3263598326359833, | |
| "eval_accuracy": 0.8465783664459161, | |
| "eval_f1": 0.6774941995359629, | |
| "eval_loss": 0.30984166264533997, | |
| "eval_precision": 0.8248587570621468, | |
| "eval_recall": 0.5748031496062992, | |
| "eval_runtime": 52.3032, | |
| "eval_samples_per_second": 5.296, | |
| "eval_steps_per_second": 0.172, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.32761506276150626, | |
| "grad_norm": 4.471529960632324, | |
| "learning_rate": 1.7008744909285626e-05, | |
| "loss": 0.2658, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.3288702928870293, | |
| "grad_norm": 4.479955673217773, | |
| "learning_rate": 1.6977380760177467e-05, | |
| "loss": 0.3076, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.3301255230125523, | |
| "grad_norm": 3.6632466316223145, | |
| "learning_rate": 1.694588228341611e-05, | |
| "loss": 0.2387, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.3313807531380753, | |
| "grad_norm": 3.813127040863037, | |
| "learning_rate": 1.691425008540625e-05, | |
| "loss": 0.2575, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.33263598326359833, | |
| "grad_norm": 3.7820916175842285, | |
| "learning_rate": 1.6882484775126968e-05, | |
| "loss": 0.2517, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.33389121338912137, | |
| "grad_norm": 3.487283229827881, | |
| "learning_rate": 1.6850586964120005e-05, | |
| "loss": 0.2898, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.33514644351464434, | |
| "grad_norm": 5.123818397521973, | |
| "learning_rate": 1.6818557266477993e-05, | |
| "loss": 0.2758, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.3364016736401674, | |
| "grad_norm": 3.208160400390625, | |
| "learning_rate": 1.6786396298832622e-05, | |
| "loss": 0.2846, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.33765690376569035, | |
| "grad_norm": 2.8521032333374023, | |
| "learning_rate": 1.6754104680342783e-05, | |
| "loss": 0.2573, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.3389121338912134, | |
| "grad_norm": 2.8169782161712646, | |
| "learning_rate": 1.6721683032682637e-05, | |
| "loss": 0.2259, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.3401673640167364, | |
| "grad_norm": 3.7779228687286377, | |
| "learning_rate": 1.6689131980029647e-05, | |
| "loss": 0.2947, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.3414225941422594, | |
| "grad_norm": 4.368408203125, | |
| "learning_rate": 1.6656452149052568e-05, | |
| "loss": 0.2654, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.3426778242677824, | |
| "grad_norm": 3.421369791030884, | |
| "learning_rate": 1.662364416889938e-05, | |
| "loss": 0.2921, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.34393305439330546, | |
| "grad_norm": 4.275522232055664, | |
| "learning_rate": 1.6590708671185176e-05, | |
| "loss": 0.2527, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.34518828451882844, | |
| "grad_norm": 3.0027596950531006, | |
| "learning_rate": 1.6557646289979996e-05, | |
| "loss": 0.2031, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.34644351464435147, | |
| "grad_norm": 3.2799339294433594, | |
| "learning_rate": 1.6524457661796626e-05, | |
| "loss": 0.2276, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.3476987447698745, | |
| "grad_norm": 3.7090659141540527, | |
| "learning_rate": 1.6491143425578345e-05, | |
| "loss": 0.2264, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.3489539748953975, | |
| "grad_norm": 6.081251621246338, | |
| "learning_rate": 1.645770422268662e-05, | |
| "loss": 0.3315, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.3502092050209205, | |
| "grad_norm": 5.695575714111328, | |
| "learning_rate": 1.6424140696888765e-05, | |
| "loss": 0.2948, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.3514644351464435, | |
| "grad_norm": 4.191822052001953, | |
| "learning_rate": 1.639045349434554e-05, | |
| "loss": 0.2868, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.3514644351464435, | |
| "eval_accuracy": 0.8432671081677704, | |
| "eval_f1": 0.6830357142857143, | |
| "eval_loss": 0.3159337043762207, | |
| "eval_precision": 0.788659793814433, | |
| "eval_recall": 0.6023622047244095, | |
| "eval_runtime": 52.6769, | |
| "eval_samples_per_second": 5.258, | |
| "eval_steps_per_second": 0.171, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.3527196652719665, | |
| "grad_norm": 4.9057183265686035, | |
| "learning_rate": 1.6356643263598716e-05, | |
| "loss": 0.3545, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.35397489539748955, | |
| "grad_norm": 6.470303058624268, | |
| "learning_rate": 1.6322710655558577e-05, | |
| "loss": 0.3414, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.35523012552301253, | |
| "grad_norm": 3.9251017570495605, | |
| "learning_rate": 1.6288656323491415e-05, | |
| "loss": 0.2573, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.35648535564853556, | |
| "grad_norm": 4.604090213775635, | |
| "learning_rate": 1.6254480923006924e-05, | |
| "loss": 0.226, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.3577405857740586, | |
| "grad_norm": 6.23361873626709, | |
| "learning_rate": 1.6220185112045606e-05, | |
| "loss": 0.2693, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.35899581589958157, | |
| "grad_norm": 3.5196187496185303, | |
| "learning_rate": 1.6185769550866073e-05, | |
| "loss": 0.2104, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.3602510460251046, | |
| "grad_norm": 5.589550495147705, | |
| "learning_rate": 1.6151234902032374e-05, | |
| "loss": 0.3379, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.36150627615062764, | |
| "grad_norm": 3.052987813949585, | |
| "learning_rate": 1.6116581830401193e-05, | |
| "loss": 0.2646, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.3627615062761506, | |
| "grad_norm": 2.715062141418457, | |
| "learning_rate": 1.60818110031091e-05, | |
| "loss": 0.2731, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.36401673640167365, | |
| "grad_norm": 3.9851012229919434, | |
| "learning_rate": 1.6046923089559667e-05, | |
| "loss": 0.2482, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.3652719665271967, | |
| "grad_norm": 4.131580352783203, | |
| "learning_rate": 1.6011918761410596e-05, | |
| "loss": 0.2916, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.36652719665271966, | |
| "grad_norm": 5.364291667938232, | |
| "learning_rate": 1.5976798692560796e-05, | |
| "loss": 0.3029, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.3677824267782427, | |
| "grad_norm": 3.139458417892456, | |
| "learning_rate": 1.5941563559137398e-05, | |
| "loss": 0.2396, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.36903765690376567, | |
| "grad_norm": 3.1862568855285645, | |
| "learning_rate": 1.5906214039482732e-05, | |
| "loss": 0.2504, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.3702928870292887, | |
| "grad_norm": 3.489682912826538, | |
| "learning_rate": 1.5870750814141296e-05, | |
| "loss": 0.2214, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.37154811715481173, | |
| "grad_norm": 4.336936950683594, | |
| "learning_rate": 1.5835174565846624e-05, | |
| "loss": 0.3056, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.3728033472803347, | |
| "grad_norm": 3.281315803527832, | |
| "learning_rate": 1.579948597950815e-05, | |
| "loss": 0.2579, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.37405857740585774, | |
| "grad_norm": 7.08855676651001, | |
| "learning_rate": 1.576368574219804e-05, | |
| "loss": 0.295, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.37531380753138077, | |
| "grad_norm": 5.177116394042969, | |
| "learning_rate": 1.5727774543137927e-05, | |
| "loss": 0.2363, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.37656903765690375, | |
| "grad_norm": 2.4472217559814453, | |
| "learning_rate": 1.5691753073685692e-05, | |
| "loss": 0.2601, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.37656903765690375, | |
| "eval_accuracy": 0.8587196467991169, | |
| "eval_f1": 0.7387755102040816, | |
| "eval_loss": 0.31048765778541565, | |
| "eval_precision": 0.7669491525423728, | |
| "eval_recall": 0.7125984251968503, | |
| "eval_runtime": 52.6769, | |
| "eval_samples_per_second": 5.258, | |
| "eval_steps_per_second": 0.171, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.3778242677824268, | |
| "grad_norm": 7.539090156555176, | |
| "learning_rate": 1.565562202732211e-05, | |
| "loss": 0.289, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.3790794979079498, | |
| "grad_norm": 7.3726420402526855, | |
| "learning_rate": 1.561938209963753e-05, | |
| "loss": 0.2752, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.3803347280334728, | |
| "grad_norm": 5.038547515869141, | |
| "learning_rate": 1.5583033988318453e-05, | |
| "loss": 0.2419, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.3815899581589958, | |
| "grad_norm": 3.0914595127105713, | |
| "learning_rate": 1.554657839313413e-05, | |
| "loss": 0.2324, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.38284518828451886, | |
| "grad_norm": 5.068948268890381, | |
| "learning_rate": 1.5510016015923084e-05, | |
| "loss": 0.2864, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.38410041841004183, | |
| "grad_norm": 4.331803321838379, | |
| "learning_rate": 1.5473347560579576e-05, | |
| "loss": 0.2247, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.38535564853556487, | |
| "grad_norm": 4.25094747543335, | |
| "learning_rate": 1.5436573733040073e-05, | |
| "loss": 0.2025, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.38661087866108784, | |
| "grad_norm": 6.317193984985352, | |
| "learning_rate": 1.539969524126967e-05, | |
| "loss": 0.2389, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.3878661087866109, | |
| "grad_norm": 5.176138401031494, | |
| "learning_rate": 1.5362712795248423e-05, | |
| "loss": 0.2235, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.3891213389121339, | |
| "grad_norm": 4.67032527923584, | |
| "learning_rate": 1.5325627106957715e-05, | |
| "loss": 0.2004, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3903765690376569, | |
| "grad_norm": 7.408180236816406, | |
| "learning_rate": 1.5288438890366534e-05, | |
| "loss": 0.3133, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.3916317991631799, | |
| "grad_norm": 4.369890213012695, | |
| "learning_rate": 1.5251148861417733e-05, | |
| "loss": 0.2798, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.39288702928870295, | |
| "grad_norm": 6.916268348693848, | |
| "learning_rate": 1.5213757738014234e-05, | |
| "loss": 0.2518, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.3941422594142259, | |
| "grad_norm": 3.2595841884613037, | |
| "learning_rate": 1.5176266240005225e-05, | |
| "loss": 0.2666, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.39539748953974896, | |
| "grad_norm": 4.970115661621094, | |
| "learning_rate": 1.513867508917229e-05, | |
| "loss": 0.2762, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.396652719665272, | |
| "grad_norm": 3.959069013595581, | |
| "learning_rate": 1.5100985009215519e-05, | |
| "loss": 0.2324, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.39790794979079497, | |
| "grad_norm": 5.496798515319824, | |
| "learning_rate": 1.5063196725739568e-05, | |
| "loss": 0.283, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.399163179916318, | |
| "grad_norm": 4.346258640289307, | |
| "learning_rate": 1.5025310966239701e-05, | |
| "loss": 0.2182, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.400418410041841, | |
| "grad_norm": 7.267153263092041, | |
| "learning_rate": 1.4987328460087778e-05, | |
| "loss": 0.2261, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.401673640167364, | |
| "grad_norm": 4.095457077026367, | |
| "learning_rate": 1.4949249938518203e-05, | |
| "loss": 0.2597, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.401673640167364, | |
| "eval_accuracy": 0.8509933774834437, | |
| "eval_f1": 0.6867749419953596, | |
| "eval_loss": 0.31619083881378174, | |
| "eval_precision": 0.8361581920903954, | |
| "eval_recall": 0.5826771653543307, | |
| "eval_runtime": 50.8111, | |
| "eval_samples_per_second": 5.452, | |
| "eval_steps_per_second": 0.177, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.40292887029288704, | |
| "grad_norm": 4.564698696136475, | |
| "learning_rate": 1.491107613461387e-05, | |
| "loss": 0.2494, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.40418410041841, | |
| "grad_norm": 3.54681134223938, | |
| "learning_rate": 1.4872807783292027e-05, | |
| "loss": 0.2396, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.40543933054393305, | |
| "grad_norm": 3.487334966659546, | |
| "learning_rate": 1.4834445621290144e-05, | |
| "loss": 0.2264, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.4066945606694561, | |
| "grad_norm": 4.941503047943115, | |
| "learning_rate": 1.4795990387151719e-05, | |
| "loss": 0.2566, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.40794979079497906, | |
| "grad_norm": 3.7651941776275635, | |
| "learning_rate": 1.4757442821212058e-05, | |
| "loss": 0.2159, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.4092050209205021, | |
| "grad_norm": 6.6421685218811035, | |
| "learning_rate": 1.4718803665584038e-05, | |
| "loss": 0.2367, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.4104602510460251, | |
| "grad_norm": 4.226874351501465, | |
| "learning_rate": 1.4680073664143799e-05, | |
| "loss": 0.2573, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.4117154811715481, | |
| "grad_norm": 5.6968536376953125, | |
| "learning_rate": 1.464125356251644e-05, | |
| "loss": 0.3498, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.41297071129707114, | |
| "grad_norm": 5.091569900512695, | |
| "learning_rate": 1.4602344108061657e-05, | |
| "loss": 0.2999, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.41422594142259417, | |
| "grad_norm": 3.757646083831787, | |
| "learning_rate": 1.4563346049859348e-05, | |
| "loss": 0.2588, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.41548117154811715, | |
| "grad_norm": 3.553725242614746, | |
| "learning_rate": 1.4524260138695206e-05, | |
| "loss": 0.3026, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.4167364016736402, | |
| "grad_norm": 4.0715765953063965, | |
| "learning_rate": 1.4485087127046256e-05, | |
| "loss": 0.3188, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.41799163179916315, | |
| "grad_norm": 3.9009945392608643, | |
| "learning_rate": 1.4445827769066374e-05, | |
| "loss": 0.2373, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.4192468619246862, | |
| "grad_norm": 3.4119412899017334, | |
| "learning_rate": 1.4406482820571759e-05, | |
| "loss": 0.2381, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.4205020920502092, | |
| "grad_norm": 7.349539756774902, | |
| "learning_rate": 1.4367053039026392e-05, | |
| "loss": 0.2528, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.4217573221757322, | |
| "grad_norm": 3.9228568077087402, | |
| "learning_rate": 1.4327539183527447e-05, | |
| "loss": 0.249, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.42301255230125523, | |
| "grad_norm": 5.134557247161865, | |
| "learning_rate": 1.4287942014790677e-05, | |
| "loss": 0.2908, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.42426778242677826, | |
| "grad_norm": 4.12131929397583, | |
| "learning_rate": 1.4248262295135779e-05, | |
| "loss": 0.2661, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.42552301255230124, | |
| "grad_norm": 3.757857322692871, | |
| "learning_rate": 1.42085007884717e-05, | |
| "loss": 0.2448, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.42677824267782427, | |
| "grad_norm": 3.9377548694610596, | |
| "learning_rate": 1.4168658260281944e-05, | |
| "loss": 0.287, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.42677824267782427, | |
| "eval_accuracy": 0.8532008830022075, | |
| "eval_f1": 0.70509977827051, | |
| "eval_loss": 0.29967617988586426, | |
| "eval_precision": 0.8071065989847716, | |
| "eval_recall": 0.6259842519685039, | |
| "eval_runtime": 50.9044, | |
| "eval_samples_per_second": 5.442, | |
| "eval_steps_per_second": 0.177, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.4280334728033473, | |
| "grad_norm": 3.8355214595794678, | |
| "learning_rate": 1.4128735477609839e-05, | |
| "loss": 0.2409, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.4292887029288703, | |
| "grad_norm": 3.9827072620391846, | |
| "learning_rate": 1.4088733209043748e-05, | |
| "loss": 0.1978, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.4305439330543933, | |
| "grad_norm": 3.053262710571289, | |
| "learning_rate": 1.4048652224702295e-05, | |
| "loss": 0.215, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.43179916317991635, | |
| "grad_norm": 3.117565155029297, | |
| "learning_rate": 1.400849329621953e-05, | |
| "loss": 0.2652, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.4330543933054393, | |
| "grad_norm": 4.665426731109619, | |
| "learning_rate": 1.3968257196730069e-05, | |
| "loss": 0.3002, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.43430962343096235, | |
| "grad_norm": 3.6823060512542725, | |
| "learning_rate": 1.3927944700854223e-05, | |
| "loss": 0.2987, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.43556485355648533, | |
| "grad_norm": 3.018756628036499, | |
| "learning_rate": 1.388755658468307e-05, | |
| "loss": 0.2399, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.43682008368200836, | |
| "grad_norm": 2.53790283203125, | |
| "learning_rate": 1.3847093625763517e-05, | |
| "loss": 0.2733, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.4380753138075314, | |
| "grad_norm": 4.417150974273682, | |
| "learning_rate": 1.3806556603083346e-05, | |
| "loss": 0.2144, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.4393305439330544, | |
| "grad_norm": 6.118602275848389, | |
| "learning_rate": 1.3765946297056192e-05, | |
| "loss": 0.3063, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.4405857740585774, | |
| "grad_norm": 3.5751051902770996, | |
| "learning_rate": 1.3725263489506542e-05, | |
| "loss": 0.1951, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.44184100418410044, | |
| "grad_norm": 5.6558837890625, | |
| "learning_rate": 1.3684508963654667e-05, | |
| "loss": 0.3366, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.4430962343096234, | |
| "grad_norm": 3.0790345668792725, | |
| "learning_rate": 1.364368350410155e-05, | |
| "loss": 0.2517, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.44435146443514645, | |
| "grad_norm": 3.3675646781921387, | |
| "learning_rate": 1.3602787896813787e-05, | |
| "loss": 0.283, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.4456066945606695, | |
| "grad_norm": 3.162820339202881, | |
| "learning_rate": 1.356182292910844e-05, | |
| "loss": 0.2131, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.44686192468619246, | |
| "grad_norm": 2.9676196575164795, | |
| "learning_rate": 1.3520789389637898e-05, | |
| "loss": 0.2782, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.4481171548117155, | |
| "grad_norm": 5.9504008293151855, | |
| "learning_rate": 1.347968806837468e-05, | |
| "loss": 0.2663, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.44937238493723847, | |
| "grad_norm": 5.749334335327148, | |
| "learning_rate": 1.3438519756596226e-05, | |
| "loss": 0.2307, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.4506276150627615, | |
| "grad_norm": 5.305976390838623, | |
| "learning_rate": 1.339728524686968e-05, | |
| "loss": 0.2, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.45188284518828453, | |
| "grad_norm": 5.051678657531738, | |
| "learning_rate": 1.335598533303662e-05, | |
| "loss": 0.3115, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.45188284518828453, | |
| "eval_accuracy": 0.8543046357615894, | |
| "eval_f1": 0.6986301369863014, | |
| "eval_loss": 0.30281126499176025, | |
| "eval_precision": 0.8315217391304348, | |
| "eval_recall": 0.6023622047244095, | |
| "eval_runtime": 53.4812, | |
| "eval_samples_per_second": 5.179, | |
| "eval_steps_per_second": 0.168, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.4531380753138075, | |
| "grad_norm": 3.419318914413452, | |
| "learning_rate": 1.331462081019776e-05, | |
| "loss": 0.2384, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.45439330543933054, | |
| "grad_norm": 3.9998960494995117, | |
| "learning_rate": 1.327319247469768e-05, | |
| "loss": 0.2815, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.4556485355648536, | |
| "grad_norm": 3.4446206092834473, | |
| "learning_rate": 1.323170112410946e-05, | |
| "loss": 0.272, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.45690376569037655, | |
| "grad_norm": 3.6744120121002197, | |
| "learning_rate": 1.319014755721934e-05, | |
| "loss": 0.2609, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.4581589958158996, | |
| "grad_norm": 4.846432209014893, | |
| "learning_rate": 1.3148532574011342e-05, | |
| "loss": 0.288, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.4594142259414226, | |
| "grad_norm": 5.032169818878174, | |
| "learning_rate": 1.3106856975651866e-05, | |
| "loss": 0.2614, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.4606694560669456, | |
| "grad_norm": 3.7232418060302734, | |
| "learning_rate": 1.3065121564474268e-05, | |
| "loss": 0.2498, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.4619246861924686, | |
| "grad_norm": 3.13726544380188, | |
| "learning_rate": 1.3023327143963415e-05, | |
| "loss": 0.2192, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.46317991631799166, | |
| "grad_norm": 5.035037994384766, | |
| "learning_rate": 1.2981474518740217e-05, | |
| "loss": 0.2971, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.46443514644351463, | |
| "grad_norm": 3.5766642093658447, | |
| "learning_rate": 1.293956449454612e-05, | |
| "loss": 0.3288, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.46569037656903767, | |
| "grad_norm": 2.6294803619384766, | |
| "learning_rate": 1.2897597878227624e-05, | |
| "loss": 0.236, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.46694560669456064, | |
| "grad_norm": 5.947935104370117, | |
| "learning_rate": 1.285557547772072e-05, | |
| "loss": 0.2743, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.4682008368200837, | |
| "grad_norm": 2.6133997440338135, | |
| "learning_rate": 1.2813498102035357e-05, | |
| "loss": 0.243, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.4694560669456067, | |
| "grad_norm": 3.549476146697998, | |
| "learning_rate": 1.2771366561239865e-05, | |
| "loss": 0.1827, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.4707112970711297, | |
| "grad_norm": 4.550835609436035, | |
| "learning_rate": 1.2729181666445338e-05, | |
| "loss": 0.2061, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.4719665271966527, | |
| "grad_norm": 4.819687843322754, | |
| "learning_rate": 1.2686944229790044e-05, | |
| "loss": 0.2638, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.47322175732217575, | |
| "grad_norm": 3.6842753887176514, | |
| "learning_rate": 1.264465506442378e-05, | |
| "loss": 0.2583, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.47447698744769873, | |
| "grad_norm": 7.268190860748291, | |
| "learning_rate": 1.2602314984492222e-05, | |
| "loss": 0.2948, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.47573221757322176, | |
| "grad_norm": 4.938054084777832, | |
| "learning_rate": 1.2559924805121236e-05, | |
| "loss": 0.2874, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.4769874476987448, | |
| "grad_norm": 8.232144355773926, | |
| "learning_rate": 1.2517485342401201e-05, | |
| "loss": 0.2654, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.4769874476987448, | |
| "eval_accuracy": 0.8543046357615894, | |
| "eval_f1": 0.7013574660633484, | |
| "eval_loss": 0.3007500171661377, | |
| "eval_precision": 0.824468085106383, | |
| "eval_recall": 0.610236220472441, | |
| "eval_runtime": 52.7757, | |
| "eval_samples_per_second": 5.249, | |
| "eval_steps_per_second": 0.171, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.47824267782426777, | |
| "grad_norm": 4.203466892242432, | |
| "learning_rate": 1.2474997413371294e-05, | |
| "loss": 0.1688, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.4794979079497908, | |
| "grad_norm": 2.6223390102386475, | |
| "learning_rate": 1.2432461836003762e-05, | |
| "loss": 0.2678, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.48075313807531384, | |
| "grad_norm": 3.429694414138794, | |
| "learning_rate": 1.238987942918817e-05, | |
| "loss": 0.2859, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.4820083682008368, | |
| "grad_norm": 2.9112257957458496, | |
| "learning_rate": 1.2347251012715629e-05, | |
| "loss": 0.2242, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.48326359832635984, | |
| "grad_norm": 4.1868896484375, | |
| "learning_rate": 1.2304577407263032e-05, | |
| "loss": 0.2995, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.4845188284518828, | |
| "grad_norm": 3.6559159755706787, | |
| "learning_rate": 1.2261859434377245e-05, | |
| "loss": 0.2115, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.48577405857740585, | |
| "grad_norm": 4.471072673797607, | |
| "learning_rate": 1.2219097916459284e-05, | |
| "loss": 0.2012, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.4870292887029289, | |
| "grad_norm": 4.849166393280029, | |
| "learning_rate": 1.2176293676748494e-05, | |
| "loss": 0.2927, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.48828451882845186, | |
| "grad_norm": 4.391753196716309, | |
| "learning_rate": 1.2133447539306689e-05, | |
| "loss": 0.3133, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.4895397489539749, | |
| "grad_norm": 2.870288848876953, | |
| "learning_rate": 1.2090560329002294e-05, | |
| "loss": 0.212, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.49079497907949793, | |
| "grad_norm": 5.226803302764893, | |
| "learning_rate": 1.2047632871494472e-05, | |
| "loss": 0.224, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.4920502092050209, | |
| "grad_norm": 3.988142728805542, | |
| "learning_rate": 1.200466599321721e-05, | |
| "loss": 0.2408, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.49330543933054394, | |
| "grad_norm": 3.6183176040649414, | |
| "learning_rate": 1.196166052136342e-05, | |
| "loss": 0.2265, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.49456066945606697, | |
| "grad_norm": 4.849849700927734, | |
| "learning_rate": 1.1918617283869021e-05, | |
| "loss": 0.2457, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.49581589958158995, | |
| "grad_norm": 4.868863105773926, | |
| "learning_rate": 1.1875537109396978e-05, | |
| "loss": 0.2463, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.497071129707113, | |
| "grad_norm": 3.9498956203460693, | |
| "learning_rate": 1.1832420827321374e-05, | |
| "loss": 0.2655, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.49832635983263596, | |
| "grad_norm": 4.637706756591797, | |
| "learning_rate": 1.1789269267711425e-05, | |
| "loss": 0.2025, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.499581589958159, | |
| "grad_norm": 3.8872170448303223, | |
| "learning_rate": 1.1746083261315505e-05, | |
| "loss": 0.2401, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.500836820083682, | |
| "grad_norm": 3.0792970657348633, | |
| "learning_rate": 1.1702863639545157e-05, | |
| "loss": 0.2804, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.502092050209205, | |
| "grad_norm": 5.217683792114258, | |
| "learning_rate": 1.165961123445908e-05, | |
| "loss": 0.2443, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.502092050209205, | |
| "eval_accuracy": 0.8565121412803532, | |
| "eval_f1": 0.7161572052401747, | |
| "eval_loss": 0.2955167889595032, | |
| "eval_precision": 0.803921568627451, | |
| "eval_recall": 0.6456692913385826, | |
| "eval_runtime": 51.6679, | |
| "eval_samples_per_second": 5.361, | |
| "eval_steps_per_second": 0.174, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.5033472803347281, | |
| "grad_norm": 3.5723674297332764, | |
| "learning_rate": 1.1616326878747115e-05, | |
| "loss": 0.2269, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.5046025104602511, | |
| "grad_norm": 3.2594027519226074, | |
| "learning_rate": 1.1573011405714214e-05, | |
| "loss": 0.2535, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.505857740585774, | |
| "grad_norm": 4.008416652679443, | |
| "learning_rate": 1.1529665649264388e-05, | |
| "loss": 0.2676, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.507112970711297, | |
| "grad_norm": 4.401585102081299, | |
| "learning_rate": 1.1486290443884666e-05, | |
| "loss": 0.2613, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.5083682008368201, | |
| "grad_norm": 4.80834436416626, | |
| "learning_rate": 1.1442886624629035e-05, | |
| "loss": 0.3053, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.5096234309623431, | |
| "grad_norm": 4.229012489318848, | |
| "learning_rate": 1.1399455027102327e-05, | |
| "loss": 0.241, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.5108786610878661, | |
| "grad_norm": 3.3926737308502197, | |
| "learning_rate": 1.1355996487444178e-05, | |
| "loss": 0.1874, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.5121338912133891, | |
| "grad_norm": 6.564388751983643, | |
| "learning_rate": 1.131251184231291e-05, | |
| "loss": 0.2087, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.5133891213389121, | |
| "grad_norm": 4.077323913574219, | |
| "learning_rate": 1.1269001928869414e-05, | |
| "loss": 0.2787, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.5146443514644351, | |
| "grad_norm": 4.169933319091797, | |
| "learning_rate": 1.122546758476105e-05, | |
| "loss": 0.2701, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.5158995815899582, | |
| "grad_norm": 4.548664093017578, | |
| "learning_rate": 1.1181909648105511e-05, | |
| "loss": 0.2711, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.5171548117154812, | |
| "grad_norm": 7.507246971130371, | |
| "learning_rate": 1.1138328957474691e-05, | |
| "loss": 0.2761, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.5184100418410041, | |
| "grad_norm": 4.293572425842285, | |
| "learning_rate": 1.1094726351878549e-05, | |
| "loss": 0.2451, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.5196652719665272, | |
| "grad_norm": 4.692587852478027, | |
| "learning_rate": 1.1051102670748939e-05, | |
| "loss": 0.2841, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.5209205020920502, | |
| "grad_norm": 3.3623526096343994, | |
| "learning_rate": 1.1007458753923455e-05, | |
| "loss": 0.245, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.5221757322175732, | |
| "grad_norm": 4.157695770263672, | |
| "learning_rate": 1.0963795441629275e-05, | |
| "loss": 0.2193, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.5234309623430963, | |
| "grad_norm": 5.14530086517334, | |
| "learning_rate": 1.0920113574466975e-05, | |
| "loss": 0.3033, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.5246861924686193, | |
| "grad_norm": 4.993683338165283, | |
| "learning_rate": 1.0876413993394346e-05, | |
| "loss": 0.2947, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.5259414225941422, | |
| "grad_norm": 6.235597610473633, | |
| "learning_rate": 1.0832697539710197e-05, | |
| "loss": 0.2765, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.5271966527196653, | |
| "grad_norm": 3.5556070804595947, | |
| "learning_rate": 1.0788965055038179e-05, | |
| "loss": 0.2743, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.5271966527196653, | |
| "eval_accuracy": 0.8543046357615894, | |
| "eval_f1": 0.695852534562212, | |
| "eval_loss": 0.30108267068862915, | |
| "eval_precision": 0.8388888888888889, | |
| "eval_recall": 0.594488188976378, | |
| "eval_runtime": 53.2652, | |
| "eval_samples_per_second": 5.2, | |
| "eval_steps_per_second": 0.169, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.5284518828451883, | |
| "grad_norm": 3.676478624343872, | |
| "learning_rate": 1.0745217381310562e-05, | |
| "loss": 0.2493, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.5297071129707113, | |
| "grad_norm": 2.608015775680542, | |
| "learning_rate": 1.0701455360752038e-05, | |
| "loss": 0.2325, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.5309623430962344, | |
| "grad_norm": 3.949383497238159, | |
| "learning_rate": 1.0657679835863497e-05, | |
| "loss": 0.3002, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.5322175732217573, | |
| "grad_norm": 4.198700904846191, | |
| "learning_rate": 1.0613891649405816e-05, | |
| "loss": 0.1867, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.5334728033472803, | |
| "grad_norm": 4.353850364685059, | |
| "learning_rate": 1.0570091644383631e-05, | |
| "loss": 0.2605, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.5347280334728034, | |
| "grad_norm": 3.590062379837036, | |
| "learning_rate": 1.0526280664029105e-05, | |
| "loss": 0.2438, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.5359832635983264, | |
| "grad_norm": 3.3330376148223877, | |
| "learning_rate": 1.0482459551785705e-05, | |
| "loss": 0.1865, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.5372384937238494, | |
| "grad_norm": 3.8933749198913574, | |
| "learning_rate": 1.0438629151291944e-05, | |
| "loss": 0.3091, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.5384937238493723, | |
| "grad_norm": 3.451763153076172, | |
| "learning_rate": 1.0394790306365154e-05, | |
| "loss": 0.2454, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.5397489539748954, | |
| "grad_norm": 3.9449098110198975, | |
| "learning_rate": 1.0350943860985249e-05, | |
| "loss": 0.2863, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.5410041841004184, | |
| "grad_norm": 3.1234138011932373, | |
| "learning_rate": 1.0307090659278453e-05, | |
| "loss": 0.2382, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.5422594142259414, | |
| "grad_norm": 4.012730598449707, | |
| "learning_rate": 1.0263231545501068e-05, | |
| "loss": 0.1927, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.5435146443514645, | |
| "grad_norm": 3.3617918491363525, | |
| "learning_rate": 1.0219367364023216e-05, | |
| "loss": 0.2557, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.5447698744769874, | |
| "grad_norm": 3.0772593021392822, | |
| "learning_rate": 1.0175498959312572e-05, | |
| "loss": 0.2144, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.5460251046025104, | |
| "grad_norm": 3.1349897384643555, | |
| "learning_rate": 1.013162717591813e-05, | |
| "loss": 0.2583, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.5472803347280335, | |
| "grad_norm": 3.4480161666870117, | |
| "learning_rate": 1.0087752858453923e-05, | |
| "loss": 0.2406, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.5485355648535565, | |
| "grad_norm": 7.350139617919922, | |
| "learning_rate": 1.0043876851582763e-05, | |
| "loss": 0.2542, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.5497907949790795, | |
| "grad_norm": 3.588282346725464, | |
| "learning_rate": 1e-05, | |
| "loss": 0.2763, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.5510460251046025, | |
| "grad_norm": 4.360295295715332, | |
| "learning_rate": 9.956123148417239e-06, | |
| "loss": 0.3057, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.5523012552301255, | |
| "grad_norm": 2.705050468444824, | |
| "learning_rate": 9.91224714154608e-06, | |
| "loss": 0.2248, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.5523012552301255, | |
| "eval_accuracy": 0.8532008830022075, | |
| "eval_f1": 0.6928406466512702, | |
| "eval_loss": 0.3030659258365631, | |
| "eval_precision": 0.8379888268156425, | |
| "eval_recall": 0.5905511811023622, | |
| "eval_runtime": 53.3187, | |
| "eval_samples_per_second": 5.195, | |
| "eval_steps_per_second": 0.169, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.5535564853556485, | |
| "grad_norm": 8.555684089660645, | |
| "learning_rate": 9.86837282408187e-06, | |
| "loss": 0.2456, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.5548117154811716, | |
| "grad_norm": 4.036064624786377, | |
| "learning_rate": 9.82450104068743e-06, | |
| "loss": 0.2197, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.5560669456066946, | |
| "grad_norm": 3.9498164653778076, | |
| "learning_rate": 9.78063263597679e-06, | |
| "loss": 0.2475, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.5573221757322175, | |
| "grad_norm": 5.120787620544434, | |
| "learning_rate": 9.736768454498935e-06, | |
| "loss": 0.2361, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.5585774058577406, | |
| "grad_norm": 3.8310952186584473, | |
| "learning_rate": 9.692909340721549e-06, | |
| "loss": 0.1948, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.5598326359832636, | |
| "grad_norm": 7.724740982055664, | |
| "learning_rate": 9.649056139014754e-06, | |
| "loss": 0.2821, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.5610878661087866, | |
| "grad_norm": 3.3741142749786377, | |
| "learning_rate": 9.605209693634849e-06, | |
| "loss": 0.2063, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.5623430962343097, | |
| "grad_norm": 4.161517143249512, | |
| "learning_rate": 9.561370848708061e-06, | |
| "loss": 0.2638, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.5635983263598326, | |
| "grad_norm": 4.094989776611328, | |
| "learning_rate": 9.517540448214299e-06, | |
| "loss": 0.2374, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.5648535564853556, | |
| "grad_norm": 5.402541160583496, | |
| "learning_rate": 9.473719335970896e-06, | |
| "loss": 0.2456, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.5661087866108787, | |
| "grad_norm": 3.0757880210876465, | |
| "learning_rate": 9.429908355616372e-06, | |
| "loss": 0.1735, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.5673640167364017, | |
| "grad_norm": 3.2094337940216064, | |
| "learning_rate": 9.38610835059419e-06, | |
| "loss": 0.1759, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.5686192468619247, | |
| "grad_norm": 6.4844489097595215, | |
| "learning_rate": 9.342320164136506e-06, | |
| "loss": 0.274, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.5698744769874478, | |
| "grad_norm": 3.759528636932373, | |
| "learning_rate": 9.298544639247965e-06, | |
| "loss": 0.2074, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.5711297071129707, | |
| "grad_norm": 4.877233028411865, | |
| "learning_rate": 9.25478261868944e-06, | |
| "loss": 0.2442, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.5723849372384937, | |
| "grad_norm": 4.176396369934082, | |
| "learning_rate": 9.211034944961825e-06, | |
| "loss": 0.2082, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.5736401673640167, | |
| "grad_norm": 4.096301078796387, | |
| "learning_rate": 9.167302460289804e-06, | |
| "loss": 0.2556, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.5748953974895398, | |
| "grad_norm": 3.9278416633605957, | |
| "learning_rate": 9.123586006605658e-06, | |
| "loss": 0.1989, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.5761506276150627, | |
| "grad_norm": 6.082350254058838, | |
| "learning_rate": 9.079886425533026e-06, | |
| "loss": 0.3095, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.5774058577405857, | |
| "grad_norm": 4.512117862701416, | |
| "learning_rate": 9.036204558370725e-06, | |
| "loss": 0.2149, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.5774058577405857, | |
| "eval_accuracy": 0.8609271523178808, | |
| "eval_f1": 0.7307692307692307, | |
| "eval_loss": 0.28679677844047546, | |
| "eval_precision": 0.7990654205607477, | |
| "eval_recall": 0.6732283464566929, | |
| "eval_runtime": 52.9315, | |
| "eval_samples_per_second": 5.233, | |
| "eval_steps_per_second": 0.17, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.5786610878661088, | |
| "grad_norm": 5.169907569885254, | |
| "learning_rate": 8.992541246076552e-06, | |
| "loss": 0.2715, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.5799163179916318, | |
| "grad_norm": 7.113213539123535, | |
| "learning_rate": 8.948897329251066e-06, | |
| "loss": 0.2462, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.5811715481171548, | |
| "grad_norm": 6.822581768035889, | |
| "learning_rate": 8.905273648121455e-06, | |
| "loss": 0.2062, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.5824267782426779, | |
| "grad_norm": 4.880428314208984, | |
| "learning_rate": 8.861671042525312e-06, | |
| "loss": 0.2825, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.5836820083682008, | |
| "grad_norm": 4.777677536010742, | |
| "learning_rate": 8.818090351894492e-06, | |
| "loss": 0.2439, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.5849372384937238, | |
| "grad_norm": 4.444671154022217, | |
| "learning_rate": 8.774532415238954e-06, | |
| "loss": 0.2612, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.5861924686192469, | |
| "grad_norm": 4.537267208099365, | |
| "learning_rate": 8.730998071130589e-06, | |
| "loss": 0.2002, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.5874476987447699, | |
| "grad_norm": 2.443470001220703, | |
| "learning_rate": 8.68748815768709e-06, | |
| "loss": 0.205, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.5887029288702929, | |
| "grad_norm": 3.0930683612823486, | |
| "learning_rate": 8.64400351255582e-06, | |
| "loss": 0.1858, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.5899581589958159, | |
| "grad_norm": 4.16900634765625, | |
| "learning_rate": 8.600544972897678e-06, | |
| "loss": 0.2589, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.5912133891213389, | |
| "grad_norm": 3.5443623065948486, | |
| "learning_rate": 8.55711337537097e-06, | |
| "loss": 0.2245, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.5924686192468619, | |
| "grad_norm": 4.173429012298584, | |
| "learning_rate": 8.513709556115335e-06, | |
| "loss": 0.2116, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.593723849372385, | |
| "grad_norm": 4.342430114746094, | |
| "learning_rate": 8.470334350735615e-06, | |
| "loss": 0.2919, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.594979079497908, | |
| "grad_norm": 3.233147144317627, | |
| "learning_rate": 8.42698859428579e-06, | |
| "loss": 0.2359, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.5962343096234309, | |
| "grad_norm": 3.438584327697754, | |
| "learning_rate": 8.383673121252887e-06, | |
| "loss": 0.2049, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.597489539748954, | |
| "grad_norm": 6.184849739074707, | |
| "learning_rate": 8.340388765540923e-06, | |
| "loss": 0.2746, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.598744769874477, | |
| "grad_norm": 3.2504115104675293, | |
| "learning_rate": 8.297136360454844e-06, | |
| "loss": 0.2135, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 5.307207107543945, | |
| "learning_rate": 8.253916738684497e-06, | |
| "loss": 0.2292, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.6012552301255231, | |
| "grad_norm": 8.800402641296387, | |
| "learning_rate": 8.21073073228858e-06, | |
| "loss": 0.3836, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.602510460251046, | |
| "grad_norm": 3.201681613922119, | |
| "learning_rate": 8.16757917267863e-06, | |
| "loss": 0.1998, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.602510460251046, | |
| "eval_accuracy": 0.8587196467991169, | |
| "eval_f1": 0.7117117117117117, | |
| "eval_loss": 0.2975335419178009, | |
| "eval_precision": 0.8315789473684211, | |
| "eval_recall": 0.6220472440944882, | |
| "eval_runtime": 52.1942, | |
| "eval_samples_per_second": 5.307, | |
| "eval_steps_per_second": 0.172, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.603765690376569, | |
| "grad_norm": 4.214263916015625, | |
| "learning_rate": 8.124462890603027e-06, | |
| "loss": 0.2576, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.605020920502092, | |
| "grad_norm": 4.688704490661621, | |
| "learning_rate": 8.081382716130982e-06, | |
| "loss": 0.2174, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.6062761506276151, | |
| "grad_norm": 4.108887672424316, | |
| "learning_rate": 8.038339478636581e-06, | |
| "loss": 0.2709, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.6075313807531381, | |
| "grad_norm": 3.246297597885132, | |
| "learning_rate": 7.995334006782793e-06, | |
| "loss": 0.2215, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.608786610878661, | |
| "grad_norm": 4.0921454429626465, | |
| "learning_rate": 7.95236712850553e-06, | |
| "loss": 0.2766, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.6100418410041841, | |
| "grad_norm": 3.6497576236724854, | |
| "learning_rate": 7.909439670997706e-06, | |
| "loss": 0.2337, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.6112970711297071, | |
| "grad_norm": 3.3551504611968994, | |
| "learning_rate": 7.866552460693314e-06, | |
| "loss": 0.208, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.6125523012552301, | |
| "grad_norm": 4.46877384185791, | |
| "learning_rate": 7.823706323251512e-06, | |
| "loss": 0.2493, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.6138075313807532, | |
| "grad_norm": 4.38779878616333, | |
| "learning_rate": 7.78090208354072e-06, | |
| "loss": 0.2635, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.6150627615062761, | |
| "grad_norm": 5.72041130065918, | |
| "learning_rate": 7.738140565622758e-06, | |
| "loss": 0.2348, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.6163179916317991, | |
| "grad_norm": 5.17399263381958, | |
| "learning_rate": 7.69542259273697e-06, | |
| "loss": 0.2208, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.6175732217573222, | |
| "grad_norm": 3.7989094257354736, | |
| "learning_rate": 7.652748987284375e-06, | |
| "loss": 0.256, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.6188284518828452, | |
| "grad_norm": 3.5693840980529785, | |
| "learning_rate": 7.610120570811833e-06, | |
| "loss": 0.2408, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.6200836820083682, | |
| "grad_norm": 3.3571958541870117, | |
| "learning_rate": 7.567538163996237e-06, | |
| "loss": 0.2263, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.6213389121338913, | |
| "grad_norm": 5.469134330749512, | |
| "learning_rate": 7.525002586628707e-06, | |
| "loss": 0.2335, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.6225941422594142, | |
| "grad_norm": 2.846597909927368, | |
| "learning_rate": 7.4825146575988e-06, | |
| "loss": 0.204, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.6238493723849372, | |
| "grad_norm": 6.401832103729248, | |
| "learning_rate": 7.440075194878769e-06, | |
| "loss": 0.2643, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.6251046025104603, | |
| "grad_norm": 4.148714065551758, | |
| "learning_rate": 7.397685015507781e-06, | |
| "loss": 0.1882, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.6263598326359833, | |
| "grad_norm": 3.8023147583007812, | |
| "learning_rate": 7.355344935576221e-06, | |
| "loss": 0.179, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.6276150627615062, | |
| "grad_norm": 3.7073490619659424, | |
| "learning_rate": 7.313055770209961e-06, | |
| "loss": 0.2459, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.6276150627615062, | |
| "eval_accuracy": 0.8509933774834437, | |
| "eval_f1": 0.6882217090069284, | |
| "eval_loss": 0.29779428243637085, | |
| "eval_precision": 0.8324022346368715, | |
| "eval_recall": 0.5866141732283464, | |
| "eval_runtime": 52.9024, | |
| "eval_samples_per_second": 5.236, | |
| "eval_steps_per_second": 0.17, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.6288702928870293, | |
| "grad_norm": 4.248108863830566, | |
| "learning_rate": 7.270818333554665e-06, | |
| "loss": 0.2752, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.6301255230125523, | |
| "grad_norm": 3.575007677078247, | |
| "learning_rate": 7.228633438760138e-06, | |
| "loss": 0.238, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.6313807531380753, | |
| "grad_norm": 5.464937210083008, | |
| "learning_rate": 7.186501897964644e-06, | |
| "loss": 0.2215, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.6326359832635984, | |
| "grad_norm": 4.194279670715332, | |
| "learning_rate": 7.144424522279283e-06, | |
| "loss": 0.2387, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.6338912133891214, | |
| "grad_norm": 4.254185199737549, | |
| "learning_rate": 7.102402121772378e-06, | |
| "loss": 0.2539, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.6351464435146443, | |
| "grad_norm": 8.297247886657715, | |
| "learning_rate": 7.060435505453884e-06, | |
| "loss": 0.2829, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.6364016736401673, | |
| "grad_norm": 5.86175012588501, | |
| "learning_rate": 7.018525481259787e-06, | |
| "loss": 0.2731, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.6376569037656904, | |
| "grad_norm": 4.584890365600586, | |
| "learning_rate": 6.976672856036586e-06, | |
| "loss": 0.1941, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.6389121338912134, | |
| "grad_norm": 4.639788627624512, | |
| "learning_rate": 6.934878435525736e-06, | |
| "loss": 0.2143, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.6401673640167364, | |
| "grad_norm": 4.085133075714111, | |
| "learning_rate": 6.893143024348137e-06, | |
| "loss": 0.2428, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.6414225941422594, | |
| "grad_norm": 5.64658784866333, | |
| "learning_rate": 6.851467425988663e-06, | |
| "loss": 0.2014, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.6426778242677824, | |
| "grad_norm": 4.088956832885742, | |
| "learning_rate": 6.809852442780664e-06, | |
| "loss": 0.185, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.6439330543933054, | |
| "grad_norm": 5.6827921867370605, | |
| "learning_rate": 6.768298875890541e-06, | |
| "loss": 0.3019, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.6451882845188285, | |
| "grad_norm": 6.169975757598877, | |
| "learning_rate": 6.726807525302319e-06, | |
| "loss": 0.2872, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.6464435146443515, | |
| "grad_norm": 4.995835304260254, | |
| "learning_rate": 6.685379189802241e-06, | |
| "loss": 0.2611, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.6476987447698744, | |
| "grad_norm": 3.8299150466918945, | |
| "learning_rate": 6.6440146669633855e-06, | |
| "loss": 0.1959, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.6489539748953975, | |
| "grad_norm": 4.961380958557129, | |
| "learning_rate": 6.602714753130322e-06, | |
| "loss": 0.2274, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.6502092050209205, | |
| "grad_norm": 4.710041522979736, | |
| "learning_rate": 6.561480243403776e-06, | |
| "loss": 0.3025, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.6514644351464435, | |
| "grad_norm": 3.762503147125244, | |
| "learning_rate": 6.520311931625325e-06, | |
| "loss": 0.25, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.6527196652719666, | |
| "grad_norm": 3.377311944961548, | |
| "learning_rate": 6.479210610362103e-06, | |
| "loss": 0.1953, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.6527196652719666, | |
| "eval_accuracy": 0.8576158940397351, | |
| "eval_f1": 0.7020785219399538, | |
| "eval_loss": 0.29889100790023804, | |
| "eval_precision": 0.8491620111731844, | |
| "eval_recall": 0.5984251968503937, | |
| "eval_runtime": 50.9239, | |
| "eval_samples_per_second": 5.439, | |
| "eval_steps_per_second": 0.177, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.6539748953974895, | |
| "grad_norm": 4.013047218322754, | |
| "learning_rate": 6.4381770708915594e-06, | |
| "loss": 0.2467, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.6552301255230125, | |
| "grad_norm": 3.758030652999878, | |
| "learning_rate": 6.397212103186214e-06, | |
| "loss": 0.2211, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.6564853556485356, | |
| "grad_norm": 3.948408603668213, | |
| "learning_rate": 6.35631649589845e-06, | |
| "loss": 0.188, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.6577405857740586, | |
| "grad_norm": 2.879676103591919, | |
| "learning_rate": 6.315491036345338e-06, | |
| "loss": 0.2009, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.6589958158995816, | |
| "grad_norm": 4.440194606781006, | |
| "learning_rate": 6.274736510493462e-06, | |
| "loss": 0.2608, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.6602510460251046, | |
| "grad_norm": 7.7365403175354, | |
| "learning_rate": 6.23405370294381e-06, | |
| "loss": 0.2733, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.6615062761506276, | |
| "grad_norm": 4.252779960632324, | |
| "learning_rate": 6.1934433969166575e-06, | |
| "loss": 0.1904, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.6627615062761506, | |
| "grad_norm": 3.153885841369629, | |
| "learning_rate": 6.1529063742364844e-06, | |
| "loss": 0.1833, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.6640167364016737, | |
| "grad_norm": 7.1857476234436035, | |
| "learning_rate": 6.112443415316934e-06, | |
| "loss": 0.2215, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.6652719665271967, | |
| "grad_norm": 2.9891092777252197, | |
| "learning_rate": 6.072055299145778e-06, | |
| "loss": 0.2156, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.6665271966527196, | |
| "grad_norm": 3.8794708251953125, | |
| "learning_rate": 6.031742803269931e-06, | |
| "loss": 0.251, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.6677824267782427, | |
| "grad_norm": 3.384833335876465, | |
| "learning_rate": 5.991506703780475e-06, | |
| "loss": 0.218, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.6690376569037657, | |
| "grad_norm": 3.629615306854248, | |
| "learning_rate": 5.95134777529771e-06, | |
| "loss": 0.2329, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.6702928870292887, | |
| "grad_norm": 3.8525490760803223, | |
| "learning_rate": 5.911266790956258e-06, | |
| "loss": 0.229, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.6715481171548117, | |
| "grad_norm": 4.439032077789307, | |
| "learning_rate": 5.871264522390165e-06, | |
| "loss": 0.2752, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.6728033472803348, | |
| "grad_norm": 3.8713276386260986, | |
| "learning_rate": 5.831341739718055e-06, | |
| "loss": 0.2427, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.6740585774058577, | |
| "grad_norm": 3.2495763301849365, | |
| "learning_rate": 5.791499211528302e-06, | |
| "loss": 0.2424, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.6753138075313807, | |
| "grad_norm": 4.471564292907715, | |
| "learning_rate": 5.751737704864224e-06, | |
| "loss": 0.1954, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.6765690376569038, | |
| "grad_norm": 4.963108539581299, | |
| "learning_rate": 5.712057985209325e-06, | |
| "loss": 0.2393, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.6778242677824268, | |
| "grad_norm": 5.474493503570557, | |
| "learning_rate": 5.672460816472556e-06, | |
| "loss": 0.3153, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.6778242677824268, | |
| "eval_accuracy": 0.8642384105960265, | |
| "eval_f1": 0.7260579064587973, | |
| "eval_loss": 0.2864134609699249, | |
| "eval_precision": 0.8358974358974359, | |
| "eval_recall": 0.6417322834645669, | |
| "eval_runtime": 51.7722, | |
| "eval_samples_per_second": 5.35, | |
| "eval_steps_per_second": 0.174, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.6790794979079497, | |
| "grad_norm": 5.037370681762695, | |
| "learning_rate": 5.632946960973611e-06, | |
| "loss": 0.2517, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.6803347280334728, | |
| "grad_norm": 5.2865142822265625, | |
| "learning_rate": 5.5935171794282426e-06, | |
| "loss": 0.2962, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.6815899581589958, | |
| "grad_norm": 3.1313962936401367, | |
| "learning_rate": 5.554172230933628e-06, | |
| "loss": 0.1967, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.6828451882845188, | |
| "grad_norm": 4.365119934082031, | |
| "learning_rate": 5.514912872953746e-06, | |
| "loss": 0.2568, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.6841004184100419, | |
| "grad_norm": 2.9407169818878174, | |
| "learning_rate": 5.4757398613047985e-06, | |
| "loss": 0.2133, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.6853556485355649, | |
| "grad_norm": 4.410444259643555, | |
| "learning_rate": 5.436653950140657e-06, | |
| "loss": 0.2656, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.6866108786610878, | |
| "grad_norm": 5.113467216491699, | |
| "learning_rate": 5.397655891938348e-06, | |
| "loss": 0.2425, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.6878661087866109, | |
| "grad_norm": 3.6607089042663574, | |
| "learning_rate": 5.35874643748356e-06, | |
| "loss": 0.1942, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.6891213389121339, | |
| "grad_norm": 4.0803914070129395, | |
| "learning_rate": 5.3199263358562e-06, | |
| "loss": 0.1702, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.6903765690376569, | |
| "grad_norm": 7.306187629699707, | |
| "learning_rate": 5.281196334415968e-06, | |
| "loss": 0.1774, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.69163179916318, | |
| "grad_norm": 6.191274166107178, | |
| "learning_rate": 5.2425571787879455e-06, | |
| "loss": 0.2996, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.6928870292887029, | |
| "grad_norm": 4.841433048248291, | |
| "learning_rate": 5.204009612848288e-06, | |
| "loss": 0.3033, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.6941422594142259, | |
| "grad_norm": 4.5002899169921875, | |
| "learning_rate": 5.165554378709857e-06, | |
| "loss": 0.2149, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.695397489539749, | |
| "grad_norm": 3.8781685829162598, | |
| "learning_rate": 5.127192216707974e-06, | |
| "loss": 0.2828, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.696652719665272, | |
| "grad_norm": 4.068243980407715, | |
| "learning_rate": 5.088923865386133e-06, | |
| "loss": 0.1836, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.697907949790795, | |
| "grad_norm": 4.608306407928467, | |
| "learning_rate": 5.050750061481799e-06, | |
| "loss": 0.2514, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.699163179916318, | |
| "grad_norm": 5.066010475158691, | |
| "learning_rate": 5.012671539912226e-06, | |
| "loss": 0.2174, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.700418410041841, | |
| "grad_norm": 4.110201358795166, | |
| "learning_rate": 4.9746890337603005e-06, | |
| "loss": 0.1672, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.701673640167364, | |
| "grad_norm": 4.265486717224121, | |
| "learning_rate": 4.936803274260434e-06, | |
| "loss": 0.2198, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.702928870292887, | |
| "grad_norm": 4.03239107131958, | |
| "learning_rate": 4.899014990784485e-06, | |
| "loss": 0.2172, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.702928870292887, | |
| "eval_accuracy": 0.8443708609271523, | |
| "eval_f1": 0.6483790523690773, | |
| "eval_loss": 0.31904953718185425, | |
| "eval_precision": 0.8843537414965986, | |
| "eval_recall": 0.5118110236220472, | |
| "eval_runtime": 51.3743, | |
| "eval_samples_per_second": 5.392, | |
| "eval_steps_per_second": 0.175, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.7041841004184101, | |
| "grad_norm": 4.523290157318115, | |
| "learning_rate": 4.861324910827714e-06, | |
| "loss": 0.2345, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.705439330543933, | |
| "grad_norm": 4.160706520080566, | |
| "learning_rate": 4.8237337599947795e-06, | |
| "loss": 0.2406, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.706694560669456, | |
| "grad_norm": 6.3733811378479, | |
| "learning_rate": 4.786242261985772e-06, | |
| "loss": 0.2486, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.7079497907949791, | |
| "grad_norm": 6.861822128295898, | |
| "learning_rate": 4.748851138582269e-06, | |
| "loss": 0.2124, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.7092050209205021, | |
| "grad_norm": 4.6429829597473145, | |
| "learning_rate": 4.711561109633466e-06, | |
| "loss": 0.2569, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.7104602510460251, | |
| "grad_norm": 3.4625086784362793, | |
| "learning_rate": 4.674372893042287e-06, | |
| "loss": 0.2174, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.7117154811715481, | |
| "grad_norm": 7.073486804962158, | |
| "learning_rate": 4.63728720475158e-06, | |
| "loss": 0.2145, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.7129707112970711, | |
| "grad_norm": 4.345584869384766, | |
| "learning_rate": 4.6003047587303376e-06, | |
| "loss": 0.1827, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.7142259414225941, | |
| "grad_norm": 4.599338054656982, | |
| "learning_rate": 4.563426266959932e-06, | |
| "loss": 0.2167, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.7154811715481172, | |
| "grad_norm": 3.284950017929077, | |
| "learning_rate": 4.526652439420427e-06, | |
| "loss": 0.1409, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.7167364016736402, | |
| "grad_norm": 3.3235983848571777, | |
| "learning_rate": 4.489983984076918e-06, | |
| "loss": 0.2377, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.7179916317991631, | |
| "grad_norm": 5.1698079109191895, | |
| "learning_rate": 4.453421606865869e-06, | |
| "loss": 0.2022, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.7192468619246862, | |
| "grad_norm": 5.351980209350586, | |
| "learning_rate": 4.416966011681548e-06, | |
| "loss": 0.1903, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.7205020920502092, | |
| "grad_norm": 2.702564239501953, | |
| "learning_rate": 4.380617900362473e-06, | |
| "loss": 0.1935, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.7217573221757322, | |
| "grad_norm": 3.770988702774048, | |
| "learning_rate": 4.34437797267789e-06, | |
| "loss": 0.175, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.7230125523012553, | |
| "grad_norm": 3.6061084270477295, | |
| "learning_rate": 4.308246926314307e-06, | |
| "loss": 0.2515, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.7242677824267783, | |
| "grad_norm": 5.896265983581543, | |
| "learning_rate": 4.272225456862076e-06, | |
| "loss": 0.2694, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.7255230125523012, | |
| "grad_norm": 7.105819225311279, | |
| "learning_rate": 4.236314257801968e-06, | |
| "loss": 0.3122, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.7267782426778243, | |
| "grad_norm": 5.691869735717773, | |
| "learning_rate": 4.200514020491854e-06, | |
| "loss": 0.2672, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.7280334728033473, | |
| "grad_norm": 3.698089838027954, | |
| "learning_rate": 4.164825434153381e-06, | |
| "loss": 0.2604, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.7280334728033473, | |
| "eval_accuracy": 0.8686534216335541, | |
| "eval_f1": 0.7384615384615385, | |
| "eval_loss": 0.28295037150382996, | |
| "eval_precision": 0.835820895522388, | |
| "eval_recall": 0.6614173228346457, | |
| "eval_runtime": 53.3249, | |
| "eval_samples_per_second": 5.195, | |
| "eval_steps_per_second": 0.169, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.7292887029288703, | |
| "grad_norm": 5.295552730560303, | |
| "learning_rate": 4.129249185858704e-06, | |
| "loss": 0.2536, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.7305439330543934, | |
| "grad_norm": 3.8178629875183105, | |
| "learning_rate": 4.093785960517269e-06, | |
| "loss": 0.2233, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.7317991631799163, | |
| "grad_norm": 5.176862716674805, | |
| "learning_rate": 4.0584364408626065e-06, | |
| "loss": 0.3026, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.7330543933054393, | |
| "grad_norm": 6.326966762542725, | |
| "learning_rate": 4.0232013074392065e-06, | |
| "loss": 0.2652, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.7343096234309623, | |
| "grad_norm": 8.710590362548828, | |
| "learning_rate": 3.988081238589406e-06, | |
| "loss": 0.2439, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.7355648535564854, | |
| "grad_norm": 5.5283026695251465, | |
| "learning_rate": 3.953076910440337e-06, | |
| "loss": 0.2445, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.7368200836820084, | |
| "grad_norm": 4.787403583526611, | |
| "learning_rate": 3.918188996890903e-06, | |
| "loss": 0.2705, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.7380753138075313, | |
| "grad_norm": 6.294352054595947, | |
| "learning_rate": 3.883418169598808e-06, | |
| "loss": 0.2813, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.7393305439330544, | |
| "grad_norm": 4.5643415451049805, | |
| "learning_rate": 3.84876509796763e-06, | |
| "loss": 0.2417, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.7405857740585774, | |
| "grad_norm": 6.011057376861572, | |
| "learning_rate": 3.814230449133928e-06, | |
| "loss": 0.3062, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.7418410041841004, | |
| "grad_norm": 5.100391387939453, | |
| "learning_rate": 3.7798148879543983e-06, | |
| "loss": 0.2424, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.7430962343096235, | |
| "grad_norm": 3.619565963745117, | |
| "learning_rate": 3.745519076993078e-06, | |
| "loss": 0.262, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.7443514644351464, | |
| "grad_norm": 5.100575923919678, | |
| "learning_rate": 3.7113436765085865e-06, | |
| "loss": 0.2577, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.7456066945606694, | |
| "grad_norm": 6.600237846374512, | |
| "learning_rate": 3.6772893444414226e-06, | |
| "loss": 0.2571, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.7468619246861925, | |
| "grad_norm": 4.155444145202637, | |
| "learning_rate": 3.643356736401289e-06, | |
| "loss": 0.2558, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.7481171548117155, | |
| "grad_norm": 3.4668867588043213, | |
| "learning_rate": 3.609546505654462e-06, | |
| "loss": 0.1694, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.7493723849372385, | |
| "grad_norm": 4.315099239349365, | |
| "learning_rate": 3.5758593031112364e-06, | |
| "loss": 0.2029, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.7506276150627615, | |
| "grad_norm": 4.79595422744751, | |
| "learning_rate": 3.5422957773133804e-06, | |
| "loss": 0.2165, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.7518828451882845, | |
| "grad_norm": 3.9190430641174316, | |
| "learning_rate": 3.5088565744216574e-06, | |
| "loss": 0.2107, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.7531380753138075, | |
| "grad_norm": 4.860176086425781, | |
| "learning_rate": 3.475542338203377e-06, | |
| "loss": 0.2671, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.7531380753138075, | |
| "eval_accuracy": 0.8565121412803532, | |
| "eval_f1": 0.6976744186046512, | |
| "eval_loss": 0.29695039987564087, | |
| "eval_precision": 0.8522727272727273, | |
| "eval_recall": 0.5905511811023622, | |
| "eval_runtime": 51.86, | |
| "eval_samples_per_second": 5.341, | |
| "eval_steps_per_second": 0.174, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.7543933054393306, | |
| "grad_norm": 5.352249622344971, | |
| "learning_rate": 3.4423537100200068e-06, | |
| "loss": 0.2344, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.7556485355648536, | |
| "grad_norm": 4.358277797698975, | |
| "learning_rate": 3.4092913288148254e-06, | |
| "loss": 0.2084, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.7569037656903765, | |
| "grad_norm": 4.106777667999268, | |
| "learning_rate": 3.3763558311006207e-06, | |
| "loss": 0.2017, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.7581589958158996, | |
| "grad_norm": 6.909910678863525, | |
| "learning_rate": 3.343547850947434e-06, | |
| "loss": 0.3994, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.7594142259414226, | |
| "grad_norm": 4.24434757232666, | |
| "learning_rate": 3.310868019970356e-06, | |
| "loss": 0.2443, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.7606694560669456, | |
| "grad_norm": 4.735110759735107, | |
| "learning_rate": 3.2783169673173666e-06, | |
| "loss": 0.2305, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.7619246861924687, | |
| "grad_norm": 4.206883907318115, | |
| "learning_rate": 3.2458953196572194e-06, | |
| "loss": 0.2061, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.7631799163179916, | |
| "grad_norm": 2.6777400970458984, | |
| "learning_rate": 3.2136037011673803e-06, | |
| "loss": 0.1777, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.7644351464435146, | |
| "grad_norm": 3.7428863048553467, | |
| "learning_rate": 3.181442733522008e-06, | |
| "loss": 0.211, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.7656903765690377, | |
| "grad_norm": 5.144768714904785, | |
| "learning_rate": 3.149413035879996e-06, | |
| "loss": 0.2571, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.7669456066945607, | |
| "grad_norm": 6.341953754425049, | |
| "learning_rate": 3.1175152248730343e-06, | |
| "loss": 0.2649, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.7682008368200837, | |
| "grad_norm": 4.261897563934326, | |
| "learning_rate": 3.085749914593752e-06, | |
| "loss": 0.2207, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.7694560669456066, | |
| "grad_norm": 3.700777053833008, | |
| "learning_rate": 3.0541177165838954e-06, | |
| "loss": 0.2589, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.7707112970711297, | |
| "grad_norm": 5.691616535186768, | |
| "learning_rate": 3.022619239822535e-06, | |
| "loss": 0.2042, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.7719665271966527, | |
| "grad_norm": 2.532932996749878, | |
| "learning_rate": 2.9912550907143766e-06, | |
| "loss": 0.2177, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.7732217573221757, | |
| "grad_norm": 6.345351219177246, | |
| "learning_rate": 2.9600258730780564e-06, | |
| "loss": 0.2141, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.7744769874476988, | |
| "grad_norm": 4.4296793937683105, | |
| "learning_rate": 2.9289321881345257e-06, | |
| "loss": 0.2414, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.7757322175732217, | |
| "grad_norm": 5.737855911254883, | |
| "learning_rate": 2.897974634495482e-06, | |
| "loss": 0.2163, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.7769874476987447, | |
| "grad_norm": 3.722508430480957, | |
| "learning_rate": 2.867153808151837e-06, | |
| "loss": 0.2102, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.7782426778242678, | |
| "grad_norm": 3.9203503131866455, | |
| "learning_rate": 2.8364703024622474e-06, | |
| "loss": 0.2049, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.7782426778242678, | |
| "eval_accuracy": 0.8587196467991169, | |
| "eval_f1": 0.7117117117117117, | |
| "eval_loss": 0.28623539209365845, | |
| "eval_precision": 0.8315789473684211, | |
| "eval_recall": 0.6220472440944882, | |
| "eval_runtime": 51.4017, | |
| "eval_samples_per_second": 5.389, | |
| "eval_steps_per_second": 0.175, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.7794979079497908, | |
| "grad_norm": 5.395325660705566, | |
| "learning_rate": 2.8059247081416887e-06, | |
| "loss": 0.2889, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.7807531380753138, | |
| "grad_norm": 3.7470452785491943, | |
| "learning_rate": 2.7755176132500905e-06, | |
| "loss": 0.2567, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.7820083682008369, | |
| "grad_norm": 3.5239267349243164, | |
| "learning_rate": 2.745249603180996e-06, | |
| "loss": 0.214, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.7832635983263598, | |
| "grad_norm": 3.5487923622131348, | |
| "learning_rate": 2.7151212606503164e-06, | |
| "loss": 0.2062, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.7845188284518828, | |
| "grad_norm": 3.3393709659576416, | |
| "learning_rate": 2.6851331656850966e-06, | |
| "loss": 0.1947, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.7857740585774059, | |
| "grad_norm": 2.6931920051574707, | |
| "learning_rate": 2.6552858956123485e-06, | |
| "loss": 0.2918, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.7870292887029289, | |
| "grad_norm": 3.748667001724243, | |
| "learning_rate": 2.625580025047956e-06, | |
| "loss": 0.2286, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.7882845188284519, | |
| "grad_norm": 3.2201664447784424, | |
| "learning_rate": 2.5960161258855807e-06, | |
| "loss": 0.2037, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.7895397489539749, | |
| "grad_norm": 3.861560821533203, | |
| "learning_rate": 2.5665947672856774e-06, | |
| "loss": 0.2023, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.7907949790794979, | |
| "grad_norm": 2.7438437938690186, | |
| "learning_rate": 2.5373165156645263e-06, | |
| "loss": 0.2018, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.7920502092050209, | |
| "grad_norm": 4.868431568145752, | |
| "learning_rate": 2.5081819346833293e-06, | |
| "loss": 0.2575, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.793305439330544, | |
| "grad_norm": 4.836085796356201, | |
| "learning_rate": 2.4791915852373604e-06, | |
| "loss": 0.206, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.794560669456067, | |
| "grad_norm": 3.095149278640747, | |
| "learning_rate": 2.450346025445165e-06, | |
| "loss": 0.173, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.7958158995815899, | |
| "grad_norm": 3.0923187732696533, | |
| "learning_rate": 2.4216458106378194e-06, | |
| "loss": 0.2016, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.797071129707113, | |
| "grad_norm": 5.667265892028809, | |
| "learning_rate": 2.3930914933482317e-06, | |
| "loss": 0.3038, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.798326359832636, | |
| "grad_norm": 4.840297222137451, | |
| "learning_rate": 2.3646836233005133e-06, | |
| "loss": 0.2346, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.799581589958159, | |
| "grad_norm": 3.6551482677459717, | |
| "learning_rate": 2.3364227473993885e-06, | |
| "loss": 0.2153, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.800836820083682, | |
| "grad_norm": 6.236753940582275, | |
| "learning_rate": 2.3083094097196766e-06, | |
| "loss": 0.3236, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.802092050209205, | |
| "grad_norm": 3.9249565601348877, | |
| "learning_rate": 2.280344151495799e-06, | |
| "loss": 0.1873, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.803347280334728, | |
| "grad_norm": 5.489711284637451, | |
| "learning_rate": 2.252527511111381e-06, | |
| "loss": 0.2972, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.803347280334728, | |
| "eval_accuracy": 0.8609271523178808, | |
| "eval_f1": 0.7149321266968326, | |
| "eval_loss": 0.2890259623527527, | |
| "eval_precision": 0.8404255319148937, | |
| "eval_recall": 0.6220472440944882, | |
| "eval_runtime": 52.9255, | |
| "eval_samples_per_second": 5.234, | |
| "eval_steps_per_second": 0.17, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.804602510460251, | |
| "grad_norm": 4.235107421875, | |
| "learning_rate": 2.224860024088863e-06, | |
| "loss": 0.2522, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.8058577405857741, | |
| "grad_norm": 5.042692184448242, | |
| "learning_rate": 2.197342223079212e-06, | |
| "loss": 0.2259, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.8071129707112971, | |
| "grad_norm": 6.554571628570557, | |
| "learning_rate": 2.1699746378516618e-06, | |
| "loss": 0.3102, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.80836820083682, | |
| "grad_norm": 5.590906143188477, | |
| "learning_rate": 2.1427577952835044e-06, | |
| "loss": 0.242, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.8096234309623431, | |
| "grad_norm": 3.1682546138763428, | |
| "learning_rate": 2.1156922193499573e-06, | |
| "loss": 0.2016, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.8108786610878661, | |
| "grad_norm": 5.03267765045166, | |
| "learning_rate": 2.088778431114068e-06, | |
| "loss": 0.3032, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.8121338912133891, | |
| "grad_norm": 5.085690498352051, | |
| "learning_rate": 2.06201694871669e-06, | |
| "loss": 0.2572, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.8133891213389122, | |
| "grad_norm": 6.410571098327637, | |
| "learning_rate": 2.0354082873665015e-06, | |
| "loss": 0.2776, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.8146443514644351, | |
| "grad_norm": 2.672776699066162, | |
| "learning_rate": 2.0089529593300916e-06, | |
| "loss": 0.1729, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.8158995815899581, | |
| "grad_norm": 4.959750175476074, | |
| "learning_rate": 1.9826514739220946e-06, | |
| "loss": 0.2254, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.8171548117154812, | |
| "grad_norm": 4.0036725997924805, | |
| "learning_rate": 1.956504337495384e-06, | |
| "loss": 0.1674, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.8184100418410042, | |
| "grad_norm": 4.883665561676025, | |
| "learning_rate": 1.9305120534313295e-06, | |
| "loss": 0.1647, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.8196652719665272, | |
| "grad_norm": 5.845210075378418, | |
| "learning_rate": 1.9046751221301018e-06, | |
| "loss": 0.3202, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.8209205020920503, | |
| "grad_norm": 5.12738037109375, | |
| "learning_rate": 1.8789940410010355e-06, | |
| "loss": 0.1829, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.8221757322175732, | |
| "grad_norm": 3.904860019683838, | |
| "learning_rate": 1.853469304453066e-06, | |
| "loss": 0.2275, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.8234309623430962, | |
| "grad_norm": 4.157991409301758, | |
| "learning_rate": 1.8281014038851963e-06, | |
| "loss": 0.2335, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.8246861924686193, | |
| "grad_norm": 6.301884651184082, | |
| "learning_rate": 1.802890827677045e-06, | |
| "loss": 0.265, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.8259414225941423, | |
| "grad_norm": 4.426259994506836, | |
| "learning_rate": 1.777838061179442e-06, | |
| "loss": 0.2346, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.8271966527196652, | |
| "grad_norm": 3.300267457962036, | |
| "learning_rate": 1.7529435867050771e-06, | |
| "loss": 0.2312, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.8284518828451883, | |
| "grad_norm": 3.5169765949249268, | |
| "learning_rate": 1.7282078835192362e-06, | |
| "loss": 0.1953, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.8284518828451883, | |
| "eval_accuracy": 0.8609271523178808, | |
| "eval_f1": 0.7136363636363636, | |
| "eval_loss": 0.2910914123058319, | |
| "eval_precision": 0.8440860215053764, | |
| "eval_recall": 0.6181102362204725, | |
| "eval_runtime": 52.6057, | |
| "eval_samples_per_second": 5.266, | |
| "eval_steps_per_second": 0.171, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.8297071129707113, | |
| "grad_norm": 3.4855308532714844, | |
| "learning_rate": 1.703631427830551e-06, | |
| "loss": 0.2036, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.8309623430962343, | |
| "grad_norm": 4.0676493644714355, | |
| "learning_rate": 1.679214692781842e-06, | |
| "loss": 0.2412, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.8322175732217573, | |
| "grad_norm": 4.750482559204102, | |
| "learning_rate": 1.6549581484410105e-06, | |
| "loss": 0.2193, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.8334728033472804, | |
| "grad_norm": 7.349670886993408, | |
| "learning_rate": 1.6308622617919823e-06, | |
| "loss": 0.2712, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.8347280334728033, | |
| "grad_norm": 3.89208984375, | |
| "learning_rate": 1.606927496725722e-06, | |
| "loss": 0.2761, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.8359832635983263, | |
| "grad_norm": 3.2538020610809326, | |
| "learning_rate": 1.583154314031311e-06, | |
| "loss": 0.1974, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.8372384937238494, | |
| "grad_norm": 4.290145397186279, | |
| "learning_rate": 1.5595431713870569e-06, | |
| "loss": 0.2562, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.8384937238493724, | |
| "grad_norm": 5.561316013336182, | |
| "learning_rate": 1.5360945233516933e-06, | |
| "loss": 0.2707, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.8397489539748954, | |
| "grad_norm": 4.019890308380127, | |
| "learning_rate": 1.5128088213556324e-06, | |
| "loss": 0.1786, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.8410041841004184, | |
| "grad_norm": 4.0204758644104, | |
| "learning_rate": 1.489686513692269e-06, | |
| "loss": 0.2377, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.8422594142259414, | |
| "grad_norm": 4.556098461151123, | |
| "learning_rate": 1.4667280455093468e-06, | |
| "loss": 0.2138, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.8435146443514644, | |
| "grad_norm": 5.089337348937988, | |
| "learning_rate": 1.4439338588004005e-06, | |
| "loss": 0.2272, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.8447698744769875, | |
| "grad_norm": 4.120019435882568, | |
| "learning_rate": 1.4213043923962322e-06, | |
| "loss": 0.2104, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.8460251046025105, | |
| "grad_norm": 3.5715548992156982, | |
| "learning_rate": 1.3988400819564707e-06, | |
| "loss": 0.2013, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.8472803347280334, | |
| "grad_norm": 4.313863277435303, | |
| "learning_rate": 1.3765413599611832e-06, | |
| "loss": 0.2265, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.8485355648535565, | |
| "grad_norm": 5.553587436676025, | |
| "learning_rate": 1.3544086557025493e-06, | |
| "loss": 0.2667, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.8497907949790795, | |
| "grad_norm": 4.209076404571533, | |
| "learning_rate": 1.3324423952765974e-06, | |
| "loss": 0.2375, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.8510460251046025, | |
| "grad_norm": 4.849112033843994, | |
| "learning_rate": 1.3106430015749971e-06, | |
| "loss": 0.2268, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.8523012552301256, | |
| "grad_norm": 4.489964485168457, | |
| "learning_rate": 1.2890108942769253e-06, | |
| "loss": 0.1934, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.8535564853556485, | |
| "grad_norm": 3.962709665298462, | |
| "learning_rate": 1.2675464898409772e-06, | |
| "loss": 0.24, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.8535564853556485, | |
| "eval_accuracy": 0.8653421633554084, | |
| "eval_f1": 0.7288888888888889, | |
| "eval_loss": 0.28242796659469604, | |
| "eval_precision": 0.8367346938775511, | |
| "eval_recall": 0.6456692913385826, | |
| "eval_runtime": 53.7633, | |
| "eval_samples_per_second": 5.152, | |
| "eval_steps_per_second": 0.167, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.8548117154811715, | |
| "grad_norm": 4.079026699066162, | |
| "learning_rate": 1.2462502014971579e-06, | |
| "loss": 0.2389, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.8560669456066946, | |
| "grad_norm": 5.030835151672363, | |
| "learning_rate": 1.2251224392389183e-06, | |
| "loss": 0.296, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.8573221757322176, | |
| "grad_norm": 4.880098819732666, | |
| "learning_rate": 1.2041636098152742e-06, | |
| "loss": 0.2687, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.8585774058577406, | |
| "grad_norm": 3.222585439682007, | |
| "learning_rate": 1.1833741167229584e-06, | |
| "loss": 0.2378, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.8598326359832636, | |
| "grad_norm": 4.026751518249512, | |
| "learning_rate": 1.1627543601986702e-06, | |
| "loss": 0.1571, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.8610878661087866, | |
| "grad_norm": 4.304843902587891, | |
| "learning_rate": 1.1423047372113538e-06, | |
| "loss": 0.2836, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.8623430962343096, | |
| "grad_norm": 4.018265247344971, | |
| "learning_rate": 1.1220256414545682e-06, | |
| "loss": 0.2154, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.8635983263598327, | |
| "grad_norm": 5.214784622192383, | |
| "learning_rate": 1.1019174633389073e-06, | |
| "loss": 0.235, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.8648535564853557, | |
| "grad_norm": 3.2286956310272217, | |
| "learning_rate": 1.0819805899844748e-06, | |
| "loss": 0.1933, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.8661087866108786, | |
| "grad_norm": 4.96722936630249, | |
| "learning_rate": 1.0622154052134392e-06, | |
| "loss": 0.2189, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.8673640167364016, | |
| "grad_norm": 5.077961444854736, | |
| "learning_rate": 1.042622289542642e-06, | |
| "loss": 0.1864, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.8686192468619247, | |
| "grad_norm": 4.23472785949707, | |
| "learning_rate": 1.0232016201762696e-06, | |
| "loss": 0.286, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.8698744769874477, | |
| "grad_norm": 4.877628803253174, | |
| "learning_rate": 1.0039537709985968e-06, | |
| "loss": 0.23, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.8711297071129707, | |
| "grad_norm": 5.60928201675415, | |
| "learning_rate": 9.848791125667867e-07, | |
| "loss": 0.2311, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.8723849372384938, | |
| "grad_norm": 4.585776329040527, | |
| "learning_rate": 9.659780121037533e-07, | |
| "loss": 0.2079, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.8736401673640167, | |
| "grad_norm": 4.836354732513428, | |
| "learning_rate": 9.472508334910946e-07, | |
| "loss": 0.1449, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.8748953974895397, | |
| "grad_norm": 6.011707782745361, | |
| "learning_rate": 9.286979372620885e-07, | |
| "loss": 0.2135, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.8761506276150628, | |
| "grad_norm": 5.969391822814941, | |
| "learning_rate": 9.103196805947522e-07, | |
| "loss": 0.2964, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.8774058577405858, | |
| "grad_norm": 5.757129192352295, | |
| "learning_rate": 8.92116417304958e-07, | |
| "loss": 0.2444, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.8786610878661087, | |
| "grad_norm": 4.445234298706055, | |
| "learning_rate": 8.740884978396358e-07, | |
| "loss": 0.282, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.8786610878661087, | |
| "eval_accuracy": 0.8631346578366446, | |
| "eval_f1": 0.7219730941704036, | |
| "eval_loss": 0.2860436737537384, | |
| "eval_precision": 0.8385416666666666, | |
| "eval_recall": 0.6338582677165354, | |
| "eval_runtime": 51.8577, | |
| "eval_samples_per_second": 5.342, | |
| "eval_steps_per_second": 0.174, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.8799163179916318, | |
| "grad_norm": 5.480141639709473, | |
| "learning_rate": 8.562362692700121e-07, | |
| "loss": 0.2179, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.8811715481171548, | |
| "grad_norm": 4.480906009674072, | |
| "learning_rate": 8.385600752849387e-07, | |
| "loss": 0.2641, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.8824267782426778, | |
| "grad_norm": 4.178032875061035, | |
| "learning_rate": 8.210602561842696e-07, | |
| "loss": 0.2121, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.8836820083682009, | |
| "grad_norm": 3.827854633331299, | |
| "learning_rate": 8.037371488723078e-07, | |
| "loss": 0.2251, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.8849372384937239, | |
| "grad_norm": 4.65687370300293, | |
| "learning_rate": 7.865910868513316e-07, | |
| "loss": 0.2668, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.8861924686192468, | |
| "grad_norm": 3.7226722240448, | |
| "learning_rate": 7.69622400215162e-07, | |
| "loss": 0.2322, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.8874476987447699, | |
| "grad_norm": 2.8721847534179688, | |
| "learning_rate": 7.528314156428085e-07, | |
| "loss": 0.2454, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.8887029288702929, | |
| "grad_norm": 6.672764301300049, | |
| "learning_rate": 7.36218456392187e-07, | |
| "loss": 0.2606, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.8899581589958159, | |
| "grad_norm": 8.37939167022705, | |
| "learning_rate": 7.197838422938908e-07, | |
| "loss": 0.1692, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.891213389121339, | |
| "grad_norm": 5.5875725746154785, | |
| "learning_rate": 7.035278897450326e-07, | |
| "loss": 0.2718, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.8924686192468619, | |
| "grad_norm": 3.284900188446045, | |
| "learning_rate": 6.874509117031636e-07, | |
| "loss": 0.2369, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.8937238493723849, | |
| "grad_norm": 3.2228903770446777, | |
| "learning_rate": 6.715532176802298e-07, | |
| "loss": 0.1676, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.894979079497908, | |
| "grad_norm": 4.319593906402588, | |
| "learning_rate": 6.558351137366292e-07, | |
| "loss": 0.2301, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.896234309623431, | |
| "grad_norm": 4.787438869476318, | |
| "learning_rate": 6.402969024753147e-07, | |
| "loss": 0.2494, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.897489539748954, | |
| "grad_norm": 5.25081205368042, | |
| "learning_rate": 6.249388830359659e-07, | |
| "loss": 0.1948, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.8987447698744769, | |
| "grad_norm": 4.489555835723877, | |
| "learning_rate": 6.097613510892364e-07, | |
| "loss": 0.3019, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 5.4843339920043945, | |
| "learning_rate": 5.947645988310524e-07, | |
| "loss": 0.2591, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.901255230125523, | |
| "grad_norm": 3.593355417251587, | |
| "learning_rate": 5.799489149769943e-07, | |
| "loss": 0.2489, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.902510460251046, | |
| "grad_norm": 4.679093837738037, | |
| "learning_rate": 5.653145847567376e-07, | |
| "loss": 0.3109, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.9037656903765691, | |
| "grad_norm": 3.6788761615753174, | |
| "learning_rate": 5.508618899085583e-07, | |
| "loss": 0.1931, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.9037656903765691, | |
| "eval_accuracy": 0.8620309050772627, | |
| "eval_f1": 0.7178329571106095, | |
| "eval_loss": 0.2884938418865204, | |
| "eval_precision": 0.8412698412698413, | |
| "eval_recall": 0.6259842519685039, | |
| "eval_runtime": 51.1882, | |
| "eval_samples_per_second": 5.411, | |
| "eval_steps_per_second": 0.176, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.905020920502092, | |
| "grad_norm": 6.400424957275391, | |
| "learning_rate": 5.365911086739117e-07, | |
| "loss": 0.2357, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.906276150627615, | |
| "grad_norm": 4.213512897491455, | |
| "learning_rate": 5.225025157920782e-07, | |
| "loss": 0.2613, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.9075313807531381, | |
| "grad_norm": 4.268280506134033, | |
| "learning_rate": 5.08596382494867e-07, | |
| "loss": 0.1831, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.9087866108786611, | |
| "grad_norm": 5.073379039764404, | |
| "learning_rate": 4.948729765014004e-07, | |
| "loss": 0.2215, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.9100418410041841, | |
| "grad_norm": 3.750990867614746, | |
| "learning_rate": 4.813325620129572e-07, | |
| "loss": 0.2024, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.9112970711297071, | |
| "grad_norm": 3.3088490962982178, | |
| "learning_rate": 4.6797539970788843e-07, | |
| "loss": 0.2044, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.9125523012552301, | |
| "grad_norm": 4.00676965713501, | |
| "learning_rate": 4.548017467365917e-07, | |
| "loss": 0.1813, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.9138075313807531, | |
| "grad_norm": 4.088128089904785, | |
| "learning_rate": 4.4181185671657634e-07, | |
| "loss": 0.2087, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.9150627615062762, | |
| "grad_norm": 2.8981070518493652, | |
| "learning_rate": 4.290059797275614e-07, | |
| "loss": 0.1629, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.9163179916317992, | |
| "grad_norm": 3.5616722106933594, | |
| "learning_rate": 4.163843623066732e-07, | |
| "loss": 0.2354, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.9175732217573221, | |
| "grad_norm": 3.7056453227996826, | |
| "learning_rate": 4.0394724744369564e-07, | |
| "loss": 0.2432, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.9188284518828452, | |
| "grad_norm": 6.088983535766602, | |
| "learning_rate": 3.916948745763938e-07, | |
| "loss": 0.1824, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.9200836820083682, | |
| "grad_norm": 4.606024742126465, | |
| "learning_rate": 3.7962747958590453e-07, | |
| "loss": 0.2111, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.9213389121338912, | |
| "grad_norm": 3.6736326217651367, | |
| "learning_rate": 3.6774529479219e-07, | |
| "loss": 0.183, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.9225941422594143, | |
| "grad_norm": 4.102712154388428, | |
| "learning_rate": 3.560485489495724e-07, | |
| "loss": 0.2266, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.9238493723849373, | |
| "grad_norm": 3.6704976558685303, | |
| "learning_rate": 3.445374672423252e-07, | |
| "loss": 0.1764, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.9251046025104602, | |
| "grad_norm": 3.65004825592041, | |
| "learning_rate": 3.3321227128033983e-07, | |
| "loss": 0.183, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.9263598326359833, | |
| "grad_norm": 5.921749591827393, | |
| "learning_rate": 3.2207317909485834e-07, | |
| "loss": 0.3058, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.9276150627615063, | |
| "grad_norm": 3.211097240447998, | |
| "learning_rate": 3.1112040513427646e-07, | |
| "loss": 0.2254, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.9288702928870293, | |
| "grad_norm": 3.559563398361206, | |
| "learning_rate": 3.0035416026001573e-07, | |
| "loss": 0.2251, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.9288702928870293, | |
| "eval_accuracy": 0.8631346578366446, | |
| "eval_f1": 0.7194570135746606, | |
| "eval_loss": 0.2898021638393402, | |
| "eval_precision": 0.8457446808510638, | |
| "eval_recall": 0.6259842519685039, | |
| "eval_runtime": 51.6937, | |
| "eval_samples_per_second": 5.358, | |
| "eval_steps_per_second": 0.174, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.9301255230125522, | |
| "grad_norm": 5.178525447845459, | |
| "learning_rate": 2.8977465174246156e-07, | |
| "loss": 0.2279, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.9313807531380753, | |
| "grad_norm": 3.763082504272461, | |
| "learning_rate": 2.7938208325697825e-07, | |
| "loss": 0.2488, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.9326359832635983, | |
| "grad_norm": 3.6214609146118164, | |
| "learning_rate": 2.691766548799779e-07, | |
| "loss": 0.2449, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.9338912133891213, | |
| "grad_norm": 4.081825256347656, | |
| "learning_rate": 2.591585630850835e-07, | |
| "loss": 0.2272, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.9351464435146444, | |
| "grad_norm": 6.377195835113525, | |
| "learning_rate": 2.493280007393295e-07, | |
| "loss": 0.2366, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.9364016736401674, | |
| "grad_norm": 6.9662981033325195, | |
| "learning_rate": 2.3968515709946296e-07, | |
| "loss": 0.2034, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.9376569037656903, | |
| "grad_norm": 2.801429510116577, | |
| "learning_rate": 2.3023021780829158e-07, | |
| "loss": 0.2079, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.9389121338912134, | |
| "grad_norm": 5.810364246368408, | |
| "learning_rate": 2.2096336489111025e-07, | |
| "loss": 0.2818, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.9401673640167364, | |
| "grad_norm": 6.120866775512695, | |
| "learning_rate": 2.1188477675220142e-07, | |
| "loss": 0.3222, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.9414225941422594, | |
| "grad_norm": 4.8275957107543945, | |
| "learning_rate": 2.0299462817139902e-07, | |
| "loss": 0.2819, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.9426778242677825, | |
| "grad_norm": 3.9053924083709717, | |
| "learning_rate": 1.942930903007212e-07, | |
| "loss": 0.1928, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.9439330543933054, | |
| "grad_norm": 3.453087091445923, | |
| "learning_rate": 1.8578033066107392e-07, | |
| "loss": 0.1989, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.9451882845188284, | |
| "grad_norm": 3.622796058654785, | |
| "learning_rate": 1.7745651313903157e-07, | |
| "loss": 0.1392, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.9464435146443515, | |
| "grad_norm": 3.5609803199768066, | |
| "learning_rate": 1.693217979836792e-07, | |
| "loss": 0.1935, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.9476987447698745, | |
| "grad_norm": 4.004082679748535, | |
| "learning_rate": 1.6137634180352303e-07, | |
| "loss": 0.2124, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.9489539748953975, | |
| "grad_norm": 3.635117292404175, | |
| "learning_rate": 1.5362029756348373e-07, | |
| "loss": 0.2254, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.9502092050209205, | |
| "grad_norm": 6.247368812561035, | |
| "learning_rate": 1.4605381458194568e-07, | |
| "loss": 0.3048, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.9514644351464435, | |
| "grad_norm": 7.196407794952393, | |
| "learning_rate": 1.386770385278835e-07, | |
| "loss": 0.2386, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.9527196652719665, | |
| "grad_norm": 4.880216121673584, | |
| "learning_rate": 1.3149011141805669e-07, | |
| "loss": 0.286, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.9539748953974896, | |
| "grad_norm": 3.848249912261963, | |
| "learning_rate": 1.2449317161427942e-07, | |
| "loss": 0.178, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.9539748953974896, | |
| "eval_accuracy": 0.8631346578366446, | |
| "eval_f1": 0.7194570135746606, | |
| "eval_loss": 0.2888832986354828, | |
| "eval_precision": 0.8457446808510638, | |
| "eval_recall": 0.6259842519685039, | |
| "eval_runtime": 51.4816, | |
| "eval_samples_per_second": 5.381, | |
| "eval_steps_per_second": 0.175, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.9552301255230126, | |
| "grad_norm": 4.997644424438477, | |
| "learning_rate": 1.1768635382075289e-07, | |
| "loss": 0.2997, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.9564853556485355, | |
| "grad_norm": 4.796181678771973, | |
| "learning_rate": 1.1106978908147381e-07, | |
| "loss": 0.2601, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.9577405857740586, | |
| "grad_norm": 3.6467103958129883, | |
| "learning_rate": 1.0464360477771218e-07, | |
| "loss": 0.2264, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.9589958158995816, | |
| "grad_norm": 5.8963303565979, | |
| "learning_rate": 9.840792462555426e-08, | |
| "loss": 0.2699, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.9602510460251046, | |
| "grad_norm": 4.986123561859131, | |
| "learning_rate": 9.236286867352785e-08, | |
| "loss": 0.2784, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.9615062761506277, | |
| "grad_norm": 4.026951313018799, | |
| "learning_rate": 8.650855330028629e-08, | |
| "loss": 0.2525, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.9627615062761506, | |
| "grad_norm": 5.082536220550537, | |
| "learning_rate": 8.08450912123715e-08, | |
| "loss": 0.2589, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.9640167364016736, | |
| "grad_norm": 5.790295124053955, | |
| "learning_rate": 7.53725914420378e-08, | |
| "loss": 0.2811, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.9652719665271966, | |
| "grad_norm": 4.651604175567627, | |
| "learning_rate": 7.009115934516253e-08, | |
| "loss": 0.2499, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.9665271966527197, | |
| "grad_norm": 4.613510608673096, | |
| "learning_rate": 6.500089659920661e-08, | |
| "loss": 0.2495, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.9677824267782427, | |
| "grad_norm": 4.089534759521484, | |
| "learning_rate": 6.010190120126602e-08, | |
| "loss": 0.2421, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.9690376569037656, | |
| "grad_norm": 4.79866886138916, | |
| "learning_rate": 5.539426746618337e-08, | |
| "loss": 0.2493, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.9702928870292887, | |
| "grad_norm": 4.052529335021973, | |
| "learning_rate": 5.0878086024727104e-08, | |
| "loss": 0.245, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.9715481171548117, | |
| "grad_norm": 3.7741782665252686, | |
| "learning_rate": 4.6553443821852893e-08, | |
| "loss": 0.2205, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.9728033472803347, | |
| "grad_norm": 3.792848587036133, | |
| "learning_rate": 4.2420424115025004e-08, | |
| "loss": 0.2297, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.9740585774058578, | |
| "grad_norm": 4.39737606048584, | |
| "learning_rate": 3.847910647261754e-08, | |
| "loss": 0.1941, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.9753138075313807, | |
| "grad_norm": 4.301551342010498, | |
| "learning_rate": 3.472956677238015e-08, | |
| "loss": 0.2069, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.9765690376569037, | |
| "grad_norm": 3.8330047130584717, | |
| "learning_rate": 3.117187719997805e-08, | |
| "loss": 0.209, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.9778242677824268, | |
| "grad_norm": 3.5165398120880127, | |
| "learning_rate": 2.780610624760094e-08, | |
| "loss": 0.1973, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.9790794979079498, | |
| "grad_norm": 4.450705528259277, | |
| "learning_rate": 2.4632318712646264e-08, | |
| "loss": 0.2431, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.9790794979079498, | |
| "eval_accuracy": 0.8631346578366446, | |
| "eval_f1": 0.7194570135746606, | |
| "eval_loss": 0.2885710895061493, | |
| "eval_precision": 0.8457446808510638, | |
| "eval_recall": 0.6259842519685039, | |
| "eval_runtime": 51.066, | |
| "eval_samples_per_second": 5.424, | |
| "eval_steps_per_second": 0.176, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.9803347280334728, | |
| "grad_norm": 5.469013690948486, | |
| "learning_rate": 2.1650575696471332e-08, | |
| "loss": 0.2557, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.9815899581589959, | |
| "grad_norm": 3.630676031112671, | |
| "learning_rate": 1.8860934603215365e-08, | |
| "loss": 0.1707, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.9828451882845188, | |
| "grad_norm": 2.949327230453491, | |
| "learning_rate": 1.626344913869593e-08, | |
| "loss": 0.1495, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.9841004184100418, | |
| "grad_norm": 4.833507537841797, | |
| "learning_rate": 1.3858169309376446e-08, | |
| "loss": 0.2054, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.9853556485355649, | |
| "grad_norm": 4.511420249938965, | |
| "learning_rate": 1.1645141421399164e-08, | |
| "loss": 0.2058, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.9866108786610879, | |
| "grad_norm": 4.471105098724365, | |
| "learning_rate": 9.624408079699221e-09, | |
| "loss": 0.2349, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.9878661087866109, | |
| "grad_norm": 4.504687309265137, | |
| "learning_rate": 7.796008187177518e-09, | |
| "loss": 0.2564, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.9891213389121339, | |
| "grad_norm": 4.099091053009033, | |
| "learning_rate": 6.1599769439590896e-09, | |
| "loss": 0.2413, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.9903765690376569, | |
| "grad_norm": 3.7011802196502686, | |
| "learning_rate": 4.716345846711434e-09, | |
| "loss": 0.2547, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.9916317991631799, | |
| "grad_norm": 4.092798233032227, | |
| "learning_rate": 3.465142688036105e-09, | |
| "loss": 0.2754, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.992887029288703, | |
| "grad_norm": 3.2027719020843506, | |
| "learning_rate": 2.4063915559402463e-09, | |
| "loss": 0.2205, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.994142259414226, | |
| "grad_norm": 4.241121768951416, | |
| "learning_rate": 1.540112833366969e-09, | |
| "loss": 0.2476, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.9953974895397489, | |
| "grad_norm": 3.92455792427063, | |
| "learning_rate": 8.663231978034425e-10, | |
| "loss": 0.2345, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.9966527196652719, | |
| "grad_norm": 5.164183616638184, | |
| "learning_rate": 3.850356209633699e-10, | |
| "loss": 0.2975, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.997907949790795, | |
| "grad_norm": 7.26465368270874, | |
| "learning_rate": 9.625936853385753e-11, | |
| "loss": 0.3067, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.999163179916318, | |
| "grad_norm": 4.6205949783325195, | |
| "learning_rate": 0.0, | |
| "loss": 0.2553, | |
| "step": 796 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 796, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.917806194389156e+17, | |
| "train_batch_size": 6, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |