| { |
| "best_metric": 0.6742756804214223, |
| "best_model_checkpoint": "drive/MyDrive/NLP_HRAF//Models/HRAF_MultiLabel_SubClasses_Kfolds/Model_5_Roberta/Learning_Rate_2e-05_Weight_Decay_0.01_fold_1/checkpoint-12450", |
| "epoch": 15.0, |
| "eval_steps": 500, |
| "global_step": 12450, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.12048192771084337, |
| "grad_norm": 0.9642070531845093, |
| "learning_rate": 1.9839357429718877e-05, |
| "loss": 0.4237, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.24096385542168675, |
| "grad_norm": 0.7253573536872864, |
| "learning_rate": 1.967871485943775e-05, |
| "loss": 0.3297, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.3614457831325301, |
| "grad_norm": 0.9410437941551208, |
| "learning_rate": 1.951807228915663e-05, |
| "loss": 0.3214, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.4819277108433735, |
| "grad_norm": 1.9593065977096558, |
| "learning_rate": 1.9357429718875505e-05, |
| "loss": 0.304, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.6024096385542169, |
| "grad_norm": 1.892052412033081, |
| "learning_rate": 1.9196787148594377e-05, |
| "loss": 0.2951, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.7228915662650602, |
| "grad_norm": 1.4741196632385254, |
| "learning_rate": 1.9036144578313255e-05, |
| "loss": 0.2843, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.8433734939759037, |
| "grad_norm": 1.5908536911010742, |
| "learning_rate": 1.887550200803213e-05, |
| "loss": 0.278, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.963855421686747, |
| "grad_norm": 2.7788383960723877, |
| "learning_rate": 1.8714859437751005e-05, |
| "loss": 0.2809, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.37130801687763715, |
| "eval_f1": 0.5538327145664992, |
| "eval_loss": 0.25523483753204346, |
| "eval_roc_auc": 0.7095420424046738, |
| "eval_runtime": 36.0551, |
| "eval_samples_per_second": 46.013, |
| "eval_steps_per_second": 5.769, |
| "step": 830 |
| }, |
| { |
| "epoch": 1.0843373493975903, |
| "grad_norm": 2.297464370727539, |
| "learning_rate": 1.855421686746988e-05, |
| "loss": 0.2649, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.2048192771084336, |
| "grad_norm": 2.1780202388763428, |
| "learning_rate": 1.8393574297188755e-05, |
| "loss": 0.2526, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.3253012048192772, |
| "grad_norm": 2.1221420764923096, |
| "learning_rate": 1.8232931726907634e-05, |
| "loss": 0.2516, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.4457831325301205, |
| "grad_norm": 2.1369152069091797, |
| "learning_rate": 1.807228915662651e-05, |
| "loss": 0.2512, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.5662650602409638, |
| "grad_norm": 1.9198232889175415, |
| "learning_rate": 1.7911646586345384e-05, |
| "loss": 0.2635, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.6867469879518073, |
| "grad_norm": 1.4999780654907227, |
| "learning_rate": 1.775100401606426e-05, |
| "loss": 0.2429, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.8072289156626506, |
| "grad_norm": 2.2243740558624268, |
| "learning_rate": 1.7590361445783134e-05, |
| "loss": 0.2341, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.927710843373494, |
| "grad_norm": 1.936194896697998, |
| "learning_rate": 1.742971887550201e-05, |
| "loss": 0.2358, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.38396624472573837, |
| "eval_f1": 0.6305942773294204, |
| "eval_loss": 0.24461282789707184, |
| "eval_roc_auc": 0.7766385397275983, |
| "eval_runtime": 36.1223, |
| "eval_samples_per_second": 45.927, |
| "eval_steps_per_second": 5.758, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.0481927710843375, |
| "grad_norm": 1.8394912481307983, |
| "learning_rate": 1.7269076305220884e-05, |
| "loss": 0.2262, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.1686746987951806, |
| "grad_norm": 1.9453928470611572, |
| "learning_rate": 1.710843373493976e-05, |
| "loss": 0.2099, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.289156626506024, |
| "grad_norm": 3.4696056842803955, |
| "learning_rate": 1.6947791164658637e-05, |
| "loss": 0.2055, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.4096385542168672, |
| "grad_norm": 3.629636764526367, |
| "learning_rate": 1.6787148594377512e-05, |
| "loss": 0.2048, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2.5301204819277108, |
| "grad_norm": 3.143533229827881, |
| "learning_rate": 1.6626506024096387e-05, |
| "loss": 0.2149, |
| "step": 2100 |
| }, |
| { |
| "epoch": 2.6506024096385543, |
| "grad_norm": 4.122682094573975, |
| "learning_rate": 1.6465863453815262e-05, |
| "loss": 0.2027, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.7710843373493974, |
| "grad_norm": 2.8808822631835938, |
| "learning_rate": 1.6305220883534137e-05, |
| "loss": 0.2153, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.891566265060241, |
| "grad_norm": 2.86367130279541, |
| "learning_rate": 1.6144578313253015e-05, |
| "loss": 0.2038, |
| "step": 2400 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_accuracy": 0.3990355635925256, |
| "eval_f1": 0.648381788261108, |
| "eval_loss": 0.23571637272834778, |
| "eval_roc_auc": 0.7872360545001245, |
| "eval_runtime": 36.0229, |
| "eval_samples_per_second": 46.054, |
| "eval_steps_per_second": 5.774, |
| "step": 2490 |
| }, |
| { |
| "epoch": 3.0120481927710845, |
| "grad_norm": 2.759669780731201, |
| "learning_rate": 1.5983935742971887e-05, |
| "loss": 0.2018, |
| "step": 2500 |
| }, |
| { |
| "epoch": 3.1325301204819276, |
| "grad_norm": 6.346066951751709, |
| "learning_rate": 1.5823293172690762e-05, |
| "loss": 0.179, |
| "step": 2600 |
| }, |
| { |
| "epoch": 3.253012048192771, |
| "grad_norm": 2.918868064880371, |
| "learning_rate": 1.566265060240964e-05, |
| "loss": 0.1788, |
| "step": 2700 |
| }, |
| { |
| "epoch": 3.3734939759036147, |
| "grad_norm": 3.852792263031006, |
| "learning_rate": 1.5502008032128516e-05, |
| "loss": 0.1727, |
| "step": 2800 |
| }, |
| { |
| "epoch": 3.4939759036144578, |
| "grad_norm": 6.0479655265808105, |
| "learning_rate": 1.534136546184739e-05, |
| "loss": 0.1791, |
| "step": 2900 |
| }, |
| { |
| "epoch": 3.6144578313253013, |
| "grad_norm": 3.1924941539764404, |
| "learning_rate": 1.5180722891566266e-05, |
| "loss": 0.1739, |
| "step": 3000 |
| }, |
| { |
| "epoch": 3.734939759036145, |
| "grad_norm": 2.770388603210449, |
| "learning_rate": 1.5020080321285142e-05, |
| "loss": 0.1723, |
| "step": 3100 |
| }, |
| { |
| "epoch": 3.855421686746988, |
| "grad_norm": 3.522843360900879, |
| "learning_rate": 1.4859437751004017e-05, |
| "loss": 0.1664, |
| "step": 3200 |
| }, |
| { |
| "epoch": 3.9759036144578315, |
| "grad_norm": 2.9013149738311768, |
| "learning_rate": 1.4698795180722894e-05, |
| "loss": 0.1591, |
| "step": 3300 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_accuracy": 0.4213381555153707, |
| "eval_f1": 0.6570518823749766, |
| "eval_loss": 0.2389475554227829, |
| "eval_roc_auc": 0.7871399448360252, |
| "eval_runtime": 35.9942, |
| "eval_samples_per_second": 46.091, |
| "eval_steps_per_second": 5.779, |
| "step": 3320 |
| }, |
| { |
| "epoch": 4.096385542168675, |
| "grad_norm": 2.389446258544922, |
| "learning_rate": 1.4538152610441769e-05, |
| "loss": 0.1488, |
| "step": 3400 |
| }, |
| { |
| "epoch": 4.216867469879518, |
| "grad_norm": 2.0893971920013428, |
| "learning_rate": 1.4377510040160642e-05, |
| "loss": 0.1431, |
| "step": 3500 |
| }, |
| { |
| "epoch": 4.337349397590361, |
| "grad_norm": 4.6121745109558105, |
| "learning_rate": 1.4216867469879519e-05, |
| "loss": 0.1436, |
| "step": 3600 |
| }, |
| { |
| "epoch": 4.457831325301205, |
| "grad_norm": 3.5218753814697266, |
| "learning_rate": 1.4056224899598394e-05, |
| "loss": 0.152, |
| "step": 3700 |
| }, |
| { |
| "epoch": 4.578313253012048, |
| "grad_norm": 1.686698317527771, |
| "learning_rate": 1.3895582329317269e-05, |
| "loss": 0.1463, |
| "step": 3800 |
| }, |
| { |
| "epoch": 4.698795180722891, |
| "grad_norm": 3.079829692840576, |
| "learning_rate": 1.3734939759036146e-05, |
| "loss": 0.1434, |
| "step": 3900 |
| }, |
| { |
| "epoch": 4.8192771084337345, |
| "grad_norm": 3.5164265632629395, |
| "learning_rate": 1.357429718875502e-05, |
| "loss": 0.1352, |
| "step": 4000 |
| }, |
| { |
| "epoch": 4.9397590361445785, |
| "grad_norm": 4.049747467041016, |
| "learning_rate": 1.3413654618473897e-05, |
| "loss": 0.1427, |
| "step": 4100 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.4207353827606992, |
| "eval_f1": 0.6415022761760243, |
| "eval_loss": 0.2513498365879059, |
| "eval_roc_auc": 0.7761577437017827, |
| "eval_runtime": 36.1323, |
| "eval_samples_per_second": 45.915, |
| "eval_steps_per_second": 5.757, |
| "step": 4150 |
| }, |
| { |
| "epoch": 5.0602409638554215, |
| "grad_norm": 3.9669837951660156, |
| "learning_rate": 1.3253012048192772e-05, |
| "loss": 0.1295, |
| "step": 4200 |
| }, |
| { |
| "epoch": 5.180722891566265, |
| "grad_norm": 2.201209783554077, |
| "learning_rate": 1.309236947791165e-05, |
| "loss": 0.1154, |
| "step": 4300 |
| }, |
| { |
| "epoch": 5.301204819277109, |
| "grad_norm": 3.613372802734375, |
| "learning_rate": 1.2931726907630524e-05, |
| "loss": 0.1133, |
| "step": 4400 |
| }, |
| { |
| "epoch": 5.421686746987952, |
| "grad_norm": 4.806926727294922, |
| "learning_rate": 1.2771084337349398e-05, |
| "loss": 0.1173, |
| "step": 4500 |
| }, |
| { |
| "epoch": 5.542168674698795, |
| "grad_norm": 3.5737357139587402, |
| "learning_rate": 1.2610441767068273e-05, |
| "loss": 0.1202, |
| "step": 4600 |
| }, |
| { |
| "epoch": 5.662650602409639, |
| "grad_norm": 5.031768798828125, |
| "learning_rate": 1.244979919678715e-05, |
| "loss": 0.1147, |
| "step": 4700 |
| }, |
| { |
| "epoch": 5.783132530120482, |
| "grad_norm": 2.068950891494751, |
| "learning_rate": 1.2289156626506024e-05, |
| "loss": 0.1168, |
| "step": 4800 |
| }, |
| { |
| "epoch": 5.903614457831325, |
| "grad_norm": 2.5072097778320312, |
| "learning_rate": 1.2128514056224901e-05, |
| "loss": 0.1226, |
| "step": 4900 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_accuracy": 0.41350210970464135, |
| "eval_f1": 0.6666666666666665, |
| "eval_loss": 0.2606567144393921, |
| "eval_roc_auc": 0.8057414996823179, |
| "eval_runtime": 36.1583, |
| "eval_samples_per_second": 45.882, |
| "eval_steps_per_second": 5.752, |
| "step": 4980 |
| }, |
| { |
| "epoch": 6.024096385542169, |
| "grad_norm": 2.0976059436798096, |
| "learning_rate": 1.1967871485943776e-05, |
| "loss": 0.1077, |
| "step": 5000 |
| }, |
| { |
| "epoch": 6.144578313253012, |
| "grad_norm": 1.4871183633804321, |
| "learning_rate": 1.1807228915662651e-05, |
| "loss": 0.0934, |
| "step": 5100 |
| }, |
| { |
| "epoch": 6.265060240963855, |
| "grad_norm": 1.9144952297210693, |
| "learning_rate": 1.1646586345381528e-05, |
| "loss": 0.0938, |
| "step": 5200 |
| }, |
| { |
| "epoch": 6.385542168674699, |
| "grad_norm": 2.3612289428710938, |
| "learning_rate": 1.1485943775100403e-05, |
| "loss": 0.0955, |
| "step": 5300 |
| }, |
| { |
| "epoch": 6.506024096385542, |
| "grad_norm": 5.222254276275635, |
| "learning_rate": 1.132530120481928e-05, |
| "loss": 0.097, |
| "step": 5400 |
| }, |
| { |
| "epoch": 6.626506024096385, |
| "grad_norm": 5.138168811798096, |
| "learning_rate": 1.1164658634538153e-05, |
| "loss": 0.0993, |
| "step": 5500 |
| }, |
| { |
| "epoch": 6.746987951807229, |
| "grad_norm": 5.146157264709473, |
| "learning_rate": 1.1004016064257028e-05, |
| "loss": 0.0954, |
| "step": 5600 |
| }, |
| { |
| "epoch": 6.867469879518072, |
| "grad_norm": 4.897678375244141, |
| "learning_rate": 1.0843373493975904e-05, |
| "loss": 0.0985, |
| "step": 5700 |
| }, |
| { |
| "epoch": 6.9879518072289155, |
| "grad_norm": 3.3976993560791016, |
| "learning_rate": 1.068273092369478e-05, |
| "loss": 0.0857, |
| "step": 5800 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_accuracy": 0.4092827004219409, |
| "eval_f1": 0.6649736985307456, |
| "eval_loss": 0.2713634967803955, |
| "eval_roc_auc": 0.7982755374895809, |
| "eval_runtime": 36.3274, |
| "eval_samples_per_second": 45.668, |
| "eval_steps_per_second": 5.726, |
| "step": 5810 |
| }, |
| { |
| "epoch": 7.108433734939759, |
| "grad_norm": 6.454195499420166, |
| "learning_rate": 1.0522088353413654e-05, |
| "loss": 0.0817, |
| "step": 5900 |
| }, |
| { |
| "epoch": 7.228915662650603, |
| "grad_norm": 1.3181122541427612, |
| "learning_rate": 1.0361445783132531e-05, |
| "loss": 0.0816, |
| "step": 6000 |
| }, |
| { |
| "epoch": 7.349397590361446, |
| "grad_norm": 4.141767501831055, |
| "learning_rate": 1.0200803212851406e-05, |
| "loss": 0.0794, |
| "step": 6100 |
| }, |
| { |
| "epoch": 7.469879518072289, |
| "grad_norm": 1.5225648880004883, |
| "learning_rate": 1.0040160642570283e-05, |
| "loss": 0.0748, |
| "step": 6200 |
| }, |
| { |
| "epoch": 7.590361445783133, |
| "grad_norm": 4.940995693206787, |
| "learning_rate": 9.879518072289156e-06, |
| "loss": 0.0733, |
| "step": 6300 |
| }, |
| { |
| "epoch": 7.710843373493976, |
| "grad_norm": 3.411694288253784, |
| "learning_rate": 9.718875502008033e-06, |
| "loss": 0.0779, |
| "step": 6400 |
| }, |
| { |
| "epoch": 7.831325301204819, |
| "grad_norm": 4.794209957122803, |
| "learning_rate": 9.558232931726908e-06, |
| "loss": 0.0778, |
| "step": 6500 |
| }, |
| { |
| "epoch": 7.951807228915663, |
| "grad_norm": 2.7523696422576904, |
| "learning_rate": 9.397590361445785e-06, |
| "loss": 0.0714, |
| "step": 6600 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_accuracy": 0.4050632911392405, |
| "eval_f1": 0.6728395061728395, |
| "eval_loss": 0.2908540666103363, |
| "eval_roc_auc": 0.8154374909432742, |
| "eval_runtime": 36.2563, |
| "eval_samples_per_second": 45.758, |
| "eval_steps_per_second": 5.737, |
| "step": 6640 |
| }, |
| { |
| "epoch": 8.072289156626505, |
| "grad_norm": 2.963338851928711, |
| "learning_rate": 9.23694779116466e-06, |
| "loss": 0.066, |
| "step": 6700 |
| }, |
| { |
| "epoch": 8.19277108433735, |
| "grad_norm": 1.1161987781524658, |
| "learning_rate": 9.076305220883535e-06, |
| "loss": 0.0665, |
| "step": 6800 |
| }, |
| { |
| "epoch": 8.313253012048193, |
| "grad_norm": 1.9074004888534546, |
| "learning_rate": 8.91566265060241e-06, |
| "loss": 0.0603, |
| "step": 6900 |
| }, |
| { |
| "epoch": 8.433734939759036, |
| "grad_norm": 4.402090549468994, |
| "learning_rate": 8.755020080321286e-06, |
| "loss": 0.0558, |
| "step": 7000 |
| }, |
| { |
| "epoch": 8.55421686746988, |
| "grad_norm": 8.068613052368164, |
| "learning_rate": 8.594377510040161e-06, |
| "loss": 0.0596, |
| "step": 7100 |
| }, |
| { |
| "epoch": 8.674698795180722, |
| "grad_norm": 0.7083752751350403, |
| "learning_rate": 8.433734939759038e-06, |
| "loss": 0.0615, |
| "step": 7200 |
| }, |
| { |
| "epoch": 8.795180722891565, |
| "grad_norm": 2.8427162170410156, |
| "learning_rate": 8.273092369477911e-06, |
| "loss": 0.056, |
| "step": 7300 |
| }, |
| { |
| "epoch": 8.91566265060241, |
| "grad_norm": 4.300654888153076, |
| "learning_rate": 8.112449799196788e-06, |
| "loss": 0.073, |
| "step": 7400 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_accuracy": 0.41832429174201324, |
| "eval_f1": 0.6730903994393833, |
| "eval_loss": 0.29414018988609314, |
| "eval_roc_auc": 0.810645259972926, |
| "eval_runtime": 36.2468, |
| "eval_samples_per_second": 45.77, |
| "eval_steps_per_second": 5.738, |
| "step": 7470 |
| }, |
| { |
| "epoch": 9.036144578313253, |
| "grad_norm": 3.129971981048584, |
| "learning_rate": 7.951807228915663e-06, |
| "loss": 0.0563, |
| "step": 7500 |
| }, |
| { |
| "epoch": 9.156626506024097, |
| "grad_norm": 2.5316765308380127, |
| "learning_rate": 7.79116465863454e-06, |
| "loss": 0.048, |
| "step": 7600 |
| }, |
| { |
| "epoch": 9.27710843373494, |
| "grad_norm": 4.083515167236328, |
| "learning_rate": 7.630522088353415e-06, |
| "loss": 0.0502, |
| "step": 7700 |
| }, |
| { |
| "epoch": 9.397590361445783, |
| "grad_norm": 1.0832017660140991, |
| "learning_rate": 7.469879518072289e-06, |
| "loss": 0.0521, |
| "step": 7800 |
| }, |
| { |
| "epoch": 9.518072289156626, |
| "grad_norm": 4.1632304191589355, |
| "learning_rate": 7.309236947791165e-06, |
| "loss": 0.0508, |
| "step": 7900 |
| }, |
| { |
| "epoch": 9.638554216867469, |
| "grad_norm": 2.0603678226470947, |
| "learning_rate": 7.148594377510041e-06, |
| "loss": 0.053, |
| "step": 8000 |
| }, |
| { |
| "epoch": 9.759036144578314, |
| "grad_norm": 2.3865954875946045, |
| "learning_rate": 6.987951807228917e-06, |
| "loss": 0.0471, |
| "step": 8100 |
| }, |
| { |
| "epoch": 9.879518072289157, |
| "grad_norm": 3.4538087844848633, |
| "learning_rate": 6.8273092369477925e-06, |
| "loss": 0.0459, |
| "step": 8200 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 1.0237865447998047, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 0.052, |
| "step": 8300 |
| }, |
| { |
| "epoch": 10.0, |
| "eval_accuracy": 0.4141048824593128, |
| "eval_f1": 0.6719829877724616, |
| "eval_loss": 0.30549874901771545, |
| "eval_roc_auc": 0.807412395917321, |
| "eval_runtime": 36.14, |
| "eval_samples_per_second": 45.905, |
| "eval_steps_per_second": 5.755, |
| "step": 8300 |
| }, |
| { |
| "epoch": 10.120481927710843, |
| "grad_norm": 3.1426589488983154, |
| "learning_rate": 6.5060240963855425e-06, |
| "loss": 0.0382, |
| "step": 8400 |
| }, |
| { |
| "epoch": 10.240963855421686, |
| "grad_norm": 5.971590995788574, |
| "learning_rate": 6.345381526104418e-06, |
| "loss": 0.0403, |
| "step": 8500 |
| }, |
| { |
| "epoch": 10.36144578313253, |
| "grad_norm": 2.7165796756744385, |
| "learning_rate": 6.184738955823294e-06, |
| "loss": 0.0417, |
| "step": 8600 |
| }, |
| { |
| "epoch": 10.481927710843374, |
| "grad_norm": 6.249508857727051, |
| "learning_rate": 6.02409638554217e-06, |
| "loss": 0.0409, |
| "step": 8700 |
| }, |
| { |
| "epoch": 10.602409638554217, |
| "grad_norm": 1.5167735815048218, |
| "learning_rate": 5.863453815261044e-06, |
| "loss": 0.0413, |
| "step": 8800 |
| }, |
| { |
| "epoch": 10.72289156626506, |
| "grad_norm": 1.3362675905227661, |
| "learning_rate": 5.70281124497992e-06, |
| "loss": 0.0381, |
| "step": 8900 |
| }, |
| { |
| "epoch": 10.843373493975903, |
| "grad_norm": 3.719500780105591, |
| "learning_rate": 5.542168674698796e-06, |
| "loss": 0.0412, |
| "step": 9000 |
| }, |
| { |
| "epoch": 10.963855421686747, |
| "grad_norm": 4.197484493255615, |
| "learning_rate": 5.381526104417672e-06, |
| "loss": 0.0429, |
| "step": 9100 |
| }, |
| { |
| "epoch": 11.0, |
| "eval_accuracy": 0.4165159734779988, |
| "eval_f1": 0.6682226211849193, |
| "eval_loss": 0.31563234329223633, |
| "eval_roc_auc": 0.8023591454661876, |
| "eval_runtime": 36.1507, |
| "eval_samples_per_second": 45.891, |
| "eval_steps_per_second": 5.754, |
| "step": 9130 |
| }, |
| { |
| "epoch": 11.08433734939759, |
| "grad_norm": 1.9439764022827148, |
| "learning_rate": 5.220883534136547e-06, |
| "loss": 0.0345, |
| "step": 9200 |
| }, |
| { |
| "epoch": 11.204819277108435, |
| "grad_norm": 5.573112487792969, |
| "learning_rate": 5.060240963855422e-06, |
| "loss": 0.0334, |
| "step": 9300 |
| }, |
| { |
| "epoch": 11.325301204819278, |
| "grad_norm": 3.091160535812378, |
| "learning_rate": 4.899598393574298e-06, |
| "loss": 0.0331, |
| "step": 9400 |
| }, |
| { |
| "epoch": 11.44578313253012, |
| "grad_norm": 4.914794445037842, |
| "learning_rate": 4.7389558232931736e-06, |
| "loss": 0.0345, |
| "step": 9500 |
| }, |
| { |
| "epoch": 11.566265060240964, |
| "grad_norm": 1.9498165845870972, |
| "learning_rate": 4.578313253012049e-06, |
| "loss": 0.0345, |
| "step": 9600 |
| }, |
| { |
| "epoch": 11.686746987951807, |
| "grad_norm": 2.1993534564971924, |
| "learning_rate": 4.4176706827309244e-06, |
| "loss": 0.0332, |
| "step": 9700 |
| }, |
| { |
| "epoch": 11.80722891566265, |
| "grad_norm": 0.7553381323814392, |
| "learning_rate": 4.2570281124497995e-06, |
| "loss": 0.0323, |
| "step": 9800 |
| }, |
| { |
| "epoch": 11.927710843373493, |
| "grad_norm": 1.3014346361160278, |
| "learning_rate": 4.096385542168675e-06, |
| "loss": 0.0323, |
| "step": 9900 |
| }, |
| { |
| "epoch": 12.0, |
| "eval_accuracy": 0.41350210970464135, |
| "eval_f1": 0.6725321133204293, |
| "eval_loss": 0.3264513611793518, |
| "eval_roc_auc": 0.8093228231966124, |
| "eval_runtime": 36.2329, |
| "eval_samples_per_second": 45.787, |
| "eval_steps_per_second": 5.741, |
| "step": 9960 |
| }, |
| { |
| "epoch": 12.048192771084338, |
| "grad_norm": 2.3635246753692627, |
| "learning_rate": 3.93574297188755e-06, |
| "loss": 0.0331, |
| "step": 10000 |
| }, |
| { |
| "epoch": 12.168674698795181, |
| "grad_norm": 3.8707635402679443, |
| "learning_rate": 3.7751004016064258e-06, |
| "loss": 0.0277, |
| "step": 10100 |
| }, |
| { |
| "epoch": 12.289156626506024, |
| "grad_norm": 3.9427218437194824, |
| "learning_rate": 3.6144578313253016e-06, |
| "loss": 0.0296, |
| "step": 10200 |
| }, |
| { |
| "epoch": 12.409638554216867, |
| "grad_norm": 0.7694936990737915, |
| "learning_rate": 3.453815261044177e-06, |
| "loss": 0.0268, |
| "step": 10300 |
| }, |
| { |
| "epoch": 12.53012048192771, |
| "grad_norm": 1.952202558517456, |
| "learning_rate": 3.2931726907630525e-06, |
| "loss": 0.0276, |
| "step": 10400 |
| }, |
| { |
| "epoch": 12.650602409638553, |
| "grad_norm": 1.1884231567382812, |
| "learning_rate": 3.132530120481928e-06, |
| "loss": 0.0262, |
| "step": 10500 |
| }, |
| { |
| "epoch": 12.771084337349398, |
| "grad_norm": 0.3486195206642151, |
| "learning_rate": 2.9718875502008034e-06, |
| "loss": 0.03, |
| "step": 10600 |
| }, |
| { |
| "epoch": 12.891566265060241, |
| "grad_norm": 0.7074311971664429, |
| "learning_rate": 2.811244979919679e-06, |
| "loss": 0.0286, |
| "step": 10700 |
| }, |
| { |
| "epoch": 13.0, |
| "eval_accuracy": 0.4110910186859554, |
| "eval_f1": 0.6732949590092447, |
| "eval_loss": 0.3310515284538269, |
| "eval_roc_auc": 0.8117622251864293, |
| "eval_runtime": 36.1171, |
| "eval_samples_per_second": 45.934, |
| "eval_steps_per_second": 5.759, |
| "step": 10790 |
| }, |
| { |
| "epoch": 13.012048192771084, |
| "grad_norm": 7.378662586212158, |
| "learning_rate": 2.6506024096385547e-06, |
| "loss": 0.0271, |
| "step": 10800 |
| }, |
| { |
| "epoch": 13.132530120481928, |
| "grad_norm": 2.406675100326538, |
| "learning_rate": 2.4899598393574297e-06, |
| "loss": 0.0251, |
| "step": 10900 |
| }, |
| { |
| "epoch": 13.25301204819277, |
| "grad_norm": 3.8638405799865723, |
| "learning_rate": 2.3293172690763055e-06, |
| "loss": 0.0258, |
| "step": 11000 |
| }, |
| { |
| "epoch": 13.373493975903614, |
| "grad_norm": 2.06321120262146, |
| "learning_rate": 2.168674698795181e-06, |
| "loss": 0.0246, |
| "step": 11100 |
| }, |
| { |
| "epoch": 13.493975903614459, |
| "grad_norm": 1.5517412424087524, |
| "learning_rate": 2.0080321285140564e-06, |
| "loss": 0.0249, |
| "step": 11200 |
| }, |
| { |
| "epoch": 13.614457831325302, |
| "grad_norm": 0.4058358669281006, |
| "learning_rate": 1.8473895582329318e-06, |
| "loss": 0.0234, |
| "step": 11300 |
| }, |
| { |
| "epoch": 13.734939759036145, |
| "grad_norm": 3.0318214893341064, |
| "learning_rate": 1.6867469879518073e-06, |
| "loss": 0.0258, |
| "step": 11400 |
| }, |
| { |
| "epoch": 13.855421686746988, |
| "grad_norm": 0.8424203991889954, |
| "learning_rate": 1.526104417670683e-06, |
| "loss": 0.0269, |
| "step": 11500 |
| }, |
| { |
| "epoch": 13.975903614457831, |
| "grad_norm": 3.9194679260253906, |
| "learning_rate": 1.3654618473895584e-06, |
| "loss": 0.0234, |
| "step": 11600 |
| }, |
| { |
| "epoch": 14.0, |
| "eval_accuracy": 0.40687160940325495, |
| "eval_f1": 0.6735155841894479, |
| "eval_loss": 0.33624783158302307, |
| "eval_roc_auc": 0.8122911998969546, |
| "eval_runtime": 36.1849, |
| "eval_samples_per_second": 45.848, |
| "eval_steps_per_second": 5.748, |
| "step": 11620 |
| }, |
| { |
| "epoch": 14.096385542168674, |
| "grad_norm": 1.4352937936782837, |
| "learning_rate": 1.2048192771084338e-06, |
| "loss": 0.0209, |
| "step": 11700 |
| }, |
| { |
| "epoch": 14.216867469879517, |
| "grad_norm": 0.7450918555259705, |
| "learning_rate": 1.0441767068273092e-06, |
| "loss": 0.0218, |
| "step": 11800 |
| }, |
| { |
| "epoch": 14.337349397590362, |
| "grad_norm": 1.0368732213974, |
| "learning_rate": 8.835341365461848e-07, |
| "loss": 0.0235, |
| "step": 11900 |
| }, |
| { |
| "epoch": 14.457831325301205, |
| "grad_norm": 6.217952728271484, |
| "learning_rate": 7.228915662650602e-07, |
| "loss": 0.0228, |
| "step": 12000 |
| }, |
| { |
| "epoch": 14.578313253012048, |
| "grad_norm": 3.589872360229492, |
| "learning_rate": 5.622489959839358e-07, |
| "loss": 0.0222, |
| "step": 12100 |
| }, |
| { |
| "epoch": 14.698795180722891, |
| "grad_norm": 2.410654306411743, |
| "learning_rate": 4.0160642570281125e-07, |
| "loss": 0.0203, |
| "step": 12200 |
| }, |
| { |
| "epoch": 14.819277108433734, |
| "grad_norm": 0.5593228340148926, |
| "learning_rate": 2.409638554216868e-07, |
| "loss": 0.0246, |
| "step": 12300 |
| }, |
| { |
| "epoch": 14.939759036144578, |
| "grad_norm": 1.7539204359054565, |
| "learning_rate": 8.032128514056224e-08, |
| "loss": 0.0237, |
| "step": 12400 |
| }, |
| { |
| "epoch": 15.0, |
| "eval_accuracy": 0.4153104279686558, |
| "eval_f1": 0.6742756804214223, |
| "eval_loss": 0.3354536294937134, |
| "eval_roc_auc": 0.810820573413045, |
| "eval_runtime": 36.2022, |
| "eval_samples_per_second": 45.826, |
| "eval_steps_per_second": 5.746, |
| "step": 12450 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 12450, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 15, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.857073408472736e+16, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|