| { | |
| "best_metric": 0.9186971783638, | |
| "best_model_checkpoint": "./Finetuned_Classification_model_200k/checkpoint-500", | |
| "epoch": 0.24, | |
| "eval_steps": 500, | |
| "global_step": 1500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1e-05, | |
| "loss": 1.037, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2e-05, | |
| "loss": 0.907, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3e-05, | |
| "loss": 0.7159, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4e-05, | |
| "loss": 0.5758, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 5e-05, | |
| "loss": 0.3454, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.997326203208557e-05, | |
| "loss": 0.1851, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.994652406417113e-05, | |
| "loss": 0.2821, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9919786096256686e-05, | |
| "loss": 0.1741, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9893048128342245e-05, | |
| "loss": 0.1749, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.986631016042781e-05, | |
| "loss": 0.2407, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.983957219251337e-05, | |
| "loss": 0.2612, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.981283422459893e-05, | |
| "loss": 0.1677, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9786096256684495e-05, | |
| "loss": 0.1123, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.975935828877006e-05, | |
| "loss": 0.1368, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.973262032085561e-05, | |
| "loss": 0.1449, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.970588235294118e-05, | |
| "loss": 0.152, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.967914438502674e-05, | |
| "loss": 0.2504, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9652406417112304e-05, | |
| "loss": 0.2578, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.962566844919786e-05, | |
| "loss": 0.1289, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.959893048128342e-05, | |
| "loss": 0.1263, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.957219251336899e-05, | |
| "loss": 0.1031, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9545454545454553e-05, | |
| "loss": 0.2453, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9518716577540106e-05, | |
| "loss": 0.0675, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.949197860962567e-05, | |
| "loss": 0.1807, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.946524064171123e-05, | |
| "loss": 0.1493, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9438502673796796e-05, | |
| "loss": 0.109, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9411764705882355e-05, | |
| "loss": 0.2148, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9385026737967914e-05, | |
| "loss": 0.2574, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.935828877005348e-05, | |
| "loss": 0.2198, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.933155080213904e-05, | |
| "loss": 0.1574, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.93048128342246e-05, | |
| "loss": 0.0866, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.9278074866310164e-05, | |
| "loss": 0.1371, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.925133689839572e-05, | |
| "loss": 0.2074, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.922459893048129e-05, | |
| "loss": 0.3628, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.919786096256685e-05, | |
| "loss": 0.1888, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.917112299465241e-05, | |
| "loss": 0.0934, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.914438502673797e-05, | |
| "loss": 0.1593, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.911764705882353e-05, | |
| "loss": 0.0721, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.909090909090909e-05, | |
| "loss": 0.1139, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.906417112299466e-05, | |
| "loss": 0.1118, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9037433155080216e-05, | |
| "loss": 0.218, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.901069518716578e-05, | |
| "loss": 0.126, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.898395721925134e-05, | |
| "loss": 0.1057, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.89572192513369e-05, | |
| "loss": 0.1825, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.8930481283422465e-05, | |
| "loss": 0.1379, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.8903743315508024e-05, | |
| "loss": 0.1311, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8877005347593584e-05, | |
| "loss": 0.0722, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.885026737967915e-05, | |
| "loss": 0.1286, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.882352941176471e-05, | |
| "loss": 0.1534, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.879679144385027e-05, | |
| "loss": 0.1332, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "eval_accuracy": 0.8395015105740181, | |
| "eval_f1": 0.8361086180158485, | |
| "eval_loss": 0.9186971783638, | |
| "eval_precision": 0.8701532825604937, | |
| "eval_recall": 0.8395015105740181, | |
| "eval_runtime": 10.5139, | |
| "eval_samples_per_second": 251.857, | |
| "eval_steps_per_second": 3.995, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8770053475935826e-05, | |
| "loss": 0.1317, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.874331550802139e-05, | |
| "loss": 0.0683, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.871657754010696e-05, | |
| "loss": 0.0474, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.868983957219252e-05, | |
| "loss": 0.2302, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8663101604278076e-05, | |
| "loss": 0.1788, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.863636363636364e-05, | |
| "loss": 0.0917, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.86096256684492e-05, | |
| "loss": 0.2001, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.858288770053476e-05, | |
| "loss": 0.3409, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.855614973262032e-05, | |
| "loss": 0.2044, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8529411764705885e-05, | |
| "loss": 0.1292, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.850267379679145e-05, | |
| "loss": 0.1422, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8475935828877e-05, | |
| "loss": 0.1165, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.844919786096257e-05, | |
| "loss": 0.1115, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8422459893048135e-05, | |
| "loss": 0.0957, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8395721925133694e-05, | |
| "loss": 0.1791, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.836898395721925e-05, | |
| "loss": 0.0691, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.834224598930481e-05, | |
| "loss": 0.2829, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.831550802139038e-05, | |
| "loss": 0.2635, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.828877005347594e-05, | |
| "loss": 0.1722, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.8262032085561496e-05, | |
| "loss": 0.0974, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.823529411764706e-05, | |
| "loss": 0.0888, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.820855614973262e-05, | |
| "loss": 0.3148, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.8181818181818186e-05, | |
| "loss": 0.3556, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.8155080213903745e-05, | |
| "loss": 0.1788, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.8128342245989304e-05, | |
| "loss": 0.4376, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.810160427807487e-05, | |
| "loss": 0.2413, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.807486631016043e-05, | |
| "loss": 0.1802, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.804812834224599e-05, | |
| "loss": 0.1364, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.8021390374331554e-05, | |
| "loss": 0.0667, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.799465240641711e-05, | |
| "loss": 0.1409, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.796791443850268e-05, | |
| "loss": 0.0897, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.794117647058824e-05, | |
| "loss": 0.0871, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.79144385026738e-05, | |
| "loss": 0.0552, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.788770053475936e-05, | |
| "loss": 0.2569, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.786096256684492e-05, | |
| "loss": 0.2877, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.783422459893048e-05, | |
| "loss": 0.1365, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7807486631016047e-05, | |
| "loss": 0.1604, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7780748663101606e-05, | |
| "loss": 0.2179, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7754010695187165e-05, | |
| "loss": 0.0969, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.772727272727273e-05, | |
| "loss": 0.0953, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.770053475935829e-05, | |
| "loss": 0.0803, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.7673796791443855e-05, | |
| "loss": 0.1573, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.7647058823529414e-05, | |
| "loss": 0.1388, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.7620320855614973e-05, | |
| "loss": 0.1559, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.759358288770054e-05, | |
| "loss": 0.2411, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.75668449197861e-05, | |
| "loss": 0.1119, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.754010695187166e-05, | |
| "loss": 0.1416, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.751336898395722e-05, | |
| "loss": 0.1073, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.748663101604278e-05, | |
| "loss": 0.1799, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.745989304812835e-05, | |
| "loss": 0.127, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_accuracy": 0.847809667673716, | |
| "eval_f1": 0.8460260531471735, | |
| "eval_loss": 1.017073392868042, | |
| "eval_precision": 0.8647086198841812, | |
| "eval_recall": 0.847809667673716, | |
| "eval_runtime": 10.5379, | |
| "eval_samples_per_second": 251.284, | |
| "eval_steps_per_second": 3.986, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.74331550802139e-05, | |
| "loss": 0.1774, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.7406417112299466e-05, | |
| "loss": 0.0869, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.737967914438503e-05, | |
| "loss": 0.0686, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.735294117647059e-05, | |
| "loss": 0.124, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.732620320855615e-05, | |
| "loss": 0.0993, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.7299465240641716e-05, | |
| "loss": 0.2451, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.7272727272727275e-05, | |
| "loss": 0.1607, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.724598930481284e-05, | |
| "loss": 0.1289, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.721925133689839e-05, | |
| "loss": 0.1813, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.719251336898396e-05, | |
| "loss": 0.1628, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.7165775401069524e-05, | |
| "loss": 0.3152, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.7139037433155083e-05, | |
| "loss": 0.3835, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.711229946524064e-05, | |
| "loss": 0.1713, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.70855614973262e-05, | |
| "loss": 0.1812, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.705882352941177e-05, | |
| "loss": 0.1928, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.703208556149733e-05, | |
| "loss": 0.2963, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.7005347593582885e-05, | |
| "loss": 0.2073, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.697860962566845e-05, | |
| "loss": 0.2299, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.695187165775402e-05, | |
| "loss": 0.1832, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.6925133689839576e-05, | |
| "loss": 0.1844, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.6898395721925135e-05, | |
| "loss": 0.4959, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.6871657754010694e-05, | |
| "loss": 0.2516, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.684491978609626e-05, | |
| "loss": 0.1754, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.681818181818182e-05, | |
| "loss": 0.2554, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.679144385026738e-05, | |
| "loss": 0.1976, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.6764705882352944e-05, | |
| "loss": 0.1906, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.673796791443851e-05, | |
| "loss": 0.0806, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.671122994652407e-05, | |
| "loss": 0.1229, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.668449197860963e-05, | |
| "loss": 0.1222, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.665775401069519e-05, | |
| "loss": 0.1049, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.663101604278075e-05, | |
| "loss": 0.1548, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.660427807486631e-05, | |
| "loss": 0.1077, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.657754010695187e-05, | |
| "loss": 0.1949, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.6550802139037436e-05, | |
| "loss": 0.1349, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.6524064171123e-05, | |
| "loss": 0.223, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.6497326203208555e-05, | |
| "loss": 0.1132, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.647058823529412e-05, | |
| "loss": 0.123, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.644385026737968e-05, | |
| "loss": 0.1741, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.6417112299465245e-05, | |
| "loss": 0.0913, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.6390374331550804e-05, | |
| "loss": 0.1428, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.636363636363636e-05, | |
| "loss": 0.1171, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.633689839572193e-05, | |
| "loss": 0.1773, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.631016042780749e-05, | |
| "loss": 0.1372, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.628342245989305e-05, | |
| "loss": 0.2797, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.625668449197861e-05, | |
| "loss": 0.2813, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.622994652406417e-05, | |
| "loss": 0.2861, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.620320855614974e-05, | |
| "loss": 0.0677, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.61764705882353e-05, | |
| "loss": 0.104, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.6149732620320856e-05, | |
| "loss": 0.1533, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.612299465240642e-05, | |
| "loss": 0.1186, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_accuracy": 0.8376132930513596, | |
| "eval_f1": 0.8363352339205664, | |
| "eval_loss": 1.0860812664031982, | |
| "eval_precision": 0.8484990166173995, | |
| "eval_recall": 0.8376132930513596, | |
| "eval_runtime": 10.5226, | |
| "eval_samples_per_second": 251.648, | |
| "eval_steps_per_second": 3.991, | |
| "step": 1500 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 18750, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 1.2629444050944e+16, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |