| { | |
| "best_metric": 0.39511793851852417, | |
| "best_model_checkpoint": "./exper_batch_16_e8/checkpoint-4900", | |
| "epoch": 8.0, | |
| "global_step": 5112, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.00019960876369327075, | |
| "loss": 4.9227, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0001992175273865415, | |
| "loss": 4.809, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.0001988262910798122, | |
| "loss": 4.6411, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00019843505477308294, | |
| "loss": 4.4483, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00019804381846635368, | |
| "loss": 4.3374, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00019765258215962445, | |
| "loss": 4.2538, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019726134585289516, | |
| "loss": 4.1378, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0001968701095461659, | |
| "loss": 4.0433, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019647887323943664, | |
| "loss": 4.0425, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019608763693270738, | |
| "loss": 3.8115, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_accuracy": 0.18619892058596763, | |
| "eval_loss": 3.7948496341705322, | |
| "eval_runtime": 43.0404, | |
| "eval_samples_per_second": 60.269, | |
| "eval_steps_per_second": 7.551, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0001956964006259781, | |
| "loss": 3.8373, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019530516431924883, | |
| "loss": 3.6742, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019491392801251957, | |
| "loss": 3.581, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.0001945226917057903, | |
| "loss": 3.4363, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00019413145539906105, | |
| "loss": 3.433, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00019374021909233179, | |
| "loss": 3.3784, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00019334898278560253, | |
| "loss": 3.2506, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00019295774647887326, | |
| "loss": 3.1149, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00019256651017214398, | |
| "loss": 3.0718, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00019217527386541472, | |
| "loss": 3.1194, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "eval_accuracy": 0.32806476484194297, | |
| "eval_loss": 3.0119543075561523, | |
| "eval_runtime": 42.8934, | |
| "eval_samples_per_second": 60.476, | |
| "eval_steps_per_second": 7.577, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00019178403755868546, | |
| "loss": 2.8329, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.0001913928012519562, | |
| "loss": 2.9553, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.0001910015649452269, | |
| "loss": 2.9676, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00019061032863849765, | |
| "loss": 2.7578, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00019021909233176841, | |
| "loss": 2.6164, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00018982785602503915, | |
| "loss": 2.7734, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00018943661971830987, | |
| "loss": 2.601, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001890453834115806, | |
| "loss": 2.8081, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00018865414710485134, | |
| "loss": 2.5308, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00018826291079812208, | |
| "loss": 2.3703, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "eval_accuracy": 0.4425597532767926, | |
| "eval_loss": 2.479144811630249, | |
| "eval_runtime": 42.6319, | |
| "eval_samples_per_second": 60.846, | |
| "eval_steps_per_second": 7.623, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001878716744913928, | |
| "loss": 2.5485, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00018748043818466354, | |
| "loss": 2.4543, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00018708920187793427, | |
| "loss": 2.3969, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00018669796557120501, | |
| "loss": 2.4463, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00018630672926447575, | |
| "loss": 2.2321, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001859154929577465, | |
| "loss": 2.1456, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00018552425665101723, | |
| "loss": 2.152, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00018513302034428797, | |
| "loss": 2.1674, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00018474178403755868, | |
| "loss": 2.1503, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00018435054773082942, | |
| "loss": 2.07, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "eval_accuracy": 0.5, | |
| "eval_loss": 2.171962261199951, | |
| "eval_runtime": 43.3053, | |
| "eval_samples_per_second": 59.9, | |
| "eval_steps_per_second": 7.505, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00018395931142410016, | |
| "loss": 2.1113, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.0001835680751173709, | |
| "loss": 2.0248, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 0.00018317683881064161, | |
| "loss": 1.9278, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 0.00018278560250391235, | |
| "loss": 1.8649, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 0.00018239436619718312, | |
| "loss": 1.9462, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 0.00018200312989045386, | |
| "loss": 2.0817, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 0.00018161189358372457, | |
| "loss": 1.8968, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 0.0001812206572769953, | |
| "loss": 1.8535, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 0.00018082942097026605, | |
| "loss": 1.8609, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 0.0001804381846635368, | |
| "loss": 1.6847, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "eval_accuracy": 0.5956052428681573, | |
| "eval_loss": 1.7290887832641602, | |
| "eval_runtime": 43.1676, | |
| "eval_samples_per_second": 60.091, | |
| "eval_steps_per_second": 7.529, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 0.0001800469483568075, | |
| "loss": 1.6108, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 0.00017965571205007824, | |
| "loss": 1.6568, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 0.00017926447574334898, | |
| "loss": 1.6968, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 0.00017887323943661972, | |
| "loss": 1.5102, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 0.00017848200312989046, | |
| "loss": 1.5518, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 0.0001780907668231612, | |
| "loss": 1.6318, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 0.00017769953051643194, | |
| "loss": 1.4528, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 0.00017730829420970268, | |
| "loss": 1.5259, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 0.00017691705790297342, | |
| "loss": 1.3553, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 0.00017652582159624413, | |
| "loss": 1.3821, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "eval_accuracy": 0.6299151888974557, | |
| "eval_loss": 1.4777374267578125, | |
| "eval_runtime": 42.6804, | |
| "eval_samples_per_second": 60.777, | |
| "eval_steps_per_second": 7.615, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 0.00017613458528951487, | |
| "loss": 1.5701, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 0.0001757433489827856, | |
| "loss": 1.4866, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 0.00017535211267605635, | |
| "loss": 1.4713, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 0.0001749608763693271, | |
| "loss": 1.2937, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 0.00017456964006259783, | |
| "loss": 1.169, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 0.00017417840375586857, | |
| "loss": 1.1436, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 0.0001737871674491393, | |
| "loss": 1.1583, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 0.00017339593114241002, | |
| "loss": 0.9302, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 0.00017300469483568076, | |
| "loss": 1.113, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 0.0001726134585289515, | |
| "loss": 0.9498, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "eval_accuracy": 0.6680801850424055, | |
| "eval_loss": 1.2934883832931519, | |
| "eval_runtime": 42.6632, | |
| "eval_samples_per_second": 60.802, | |
| "eval_steps_per_second": 7.618, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 0.00017222222222222224, | |
| "loss": 0.9322, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 0.00017183098591549295, | |
| "loss": 0.9719, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 0.0001714397496087637, | |
| "loss": 0.9679, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 0.00017104851330203443, | |
| "loss": 0.8539, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 0.0001706572769953052, | |
| "loss": 0.971, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 0.0001702660406885759, | |
| "loss": 0.9654, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 0.00016987480438184665, | |
| "loss": 0.7825, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 0.00016948356807511739, | |
| "loss": 1.0415, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 0.00016909233176838813, | |
| "loss": 0.9053, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 0.00016870109546165884, | |
| "loss": 0.8741, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "eval_accuracy": 0.70508866615266, | |
| "eval_loss": 1.135292649269104, | |
| "eval_runtime": 42.8607, | |
| "eval_samples_per_second": 60.522, | |
| "eval_steps_per_second": 7.583, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 0.00016830985915492958, | |
| "loss": 0.98, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 0.00016791862284820032, | |
| "loss": 0.964, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 0.00016752738654147106, | |
| "loss": 1.0007, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 0.0001671361502347418, | |
| "loss": 0.7954, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 0.00016674491392801253, | |
| "loss": 0.9806, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 0.00016635367762128327, | |
| "loss": 0.8854, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 0.00016596244131455401, | |
| "loss": 0.9212, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 0.00016557120500782473, | |
| "loss": 0.8027, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 0.00016517996870109547, | |
| "loss": 0.7901, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 0.0001647887323943662, | |
| "loss": 0.8875, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "eval_accuracy": 0.7447956823438705, | |
| "eval_loss": 0.9951499700546265, | |
| "eval_runtime": 42.5213, | |
| "eval_samples_per_second": 61.005, | |
| "eval_steps_per_second": 7.643, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 0.00016439749608763694, | |
| "loss": 0.8624, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 0.00016400625978090766, | |
| "loss": 0.7194, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 0.0001636150234741784, | |
| "loss": 0.8012, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 0.00016322378716744916, | |
| "loss": 0.9272, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 0.0001628325508607199, | |
| "loss": 0.8623, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 0.00016244131455399061, | |
| "loss": 0.8576, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 0.00016205007824726135, | |
| "loss": 0.6199, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 0.0001616588419405321, | |
| "loss": 0.6498, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 0.00016126760563380283, | |
| "loss": 0.8226, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 0.00016087636932707354, | |
| "loss": 0.7233, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "eval_accuracy": 0.748650732459522, | |
| "eval_loss": 0.926507294178009, | |
| "eval_runtime": 43.2867, | |
| "eval_samples_per_second": 59.926, | |
| "eval_steps_per_second": 7.508, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 0.00016048513302034428, | |
| "loss": 0.7018, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 0.00016009389671361502, | |
| "loss": 0.739, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 0.00015970266040688576, | |
| "loss": 0.6795, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 0.0001593114241001565, | |
| "loss": 0.6543, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 0.00015892018779342724, | |
| "loss": 0.7367, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 0.00015852895148669798, | |
| "loss": 0.6326, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 0.00015813771517996872, | |
| "loss": 0.5459, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 0.00015774647887323943, | |
| "loss": 0.8587, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 0.00015735524256651017, | |
| "loss": 0.6657, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 0.0001569640062597809, | |
| "loss": 0.6696, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "eval_accuracy": 0.7625289128758674, | |
| "eval_loss": 0.8659528493881226, | |
| "eval_runtime": 42.5957, | |
| "eval_samples_per_second": 60.898, | |
| "eval_steps_per_second": 7.63, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 0.00015657276995305165, | |
| "loss": 0.7146, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 0.00015618153364632236, | |
| "loss": 0.4619, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 0.0001557902973395931, | |
| "loss": 0.6367, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 0.00015539906103286387, | |
| "loss": 0.585, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 0.0001550078247261346, | |
| "loss": 0.5464, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 0.00015461658841940532, | |
| "loss": 0.7695, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 0.00015422535211267606, | |
| "loss": 0.7019, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 0.0001538341158059468, | |
| "loss": 0.4508, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 0.00015344287949921754, | |
| "loss": 0.6396, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 0.00015305164319248828, | |
| "loss": 0.7364, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "eval_accuracy": 0.7579028527370856, | |
| "eval_loss": 0.8710386753082275, | |
| "eval_runtime": 42.6201, | |
| "eval_samples_per_second": 60.863, | |
| "eval_steps_per_second": 7.626, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 0.000152660406885759, | |
| "loss": 0.8301, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 0.00015226917057902973, | |
| "loss": 0.6424, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 0.00015187793427230047, | |
| "loss": 0.788, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 0.0001514866979655712, | |
| "loss": 0.8115, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 0.00015109546165884195, | |
| "loss": 0.7002, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 0.0001507042253521127, | |
| "loss": 0.5162, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 0.00015031298904538343, | |
| "loss": 0.5706, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 0.00014992175273865417, | |
| "loss": 0.4565, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 0.00014953051643192488, | |
| "loss": 0.3758, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 0.00014913928012519562, | |
| "loss": 0.3933, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "eval_accuracy": 0.8037779491133384, | |
| "eval_loss": 0.7161900997161865, | |
| "eval_runtime": 42.7187, | |
| "eval_samples_per_second": 60.723, | |
| "eval_steps_per_second": 7.608, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 0.00014874804381846636, | |
| "loss": 0.2625, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 0.0001483568075117371, | |
| "loss": 0.3585, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 0.00014796557120500784, | |
| "loss": 0.3735, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 0.00014757433489827858, | |
| "loss": 0.4221, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 0.00014718309859154932, | |
| "loss": 0.3249, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 0.00014679186228482006, | |
| "loss": 0.4802, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 0.00014640062597809077, | |
| "loss": 0.3311, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 0.0001460093896713615, | |
| "loss": 0.4276, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 0.00014561815336463225, | |
| "loss": 0.222, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 0.00014522691705790299, | |
| "loss": 0.3443, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "eval_accuracy": 0.8299922898997687, | |
| "eval_loss": 0.6305217742919922, | |
| "eval_runtime": 43.2454, | |
| "eval_samples_per_second": 59.983, | |
| "eval_steps_per_second": 7.515, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 0.0001448356807511737, | |
| "loss": 0.3609, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 0.00014444444444444444, | |
| "loss": 0.4267, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 0.00014405320813771518, | |
| "loss": 0.4206, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 0.00014366197183098594, | |
| "loss": 0.3414, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 0.00014327073552425666, | |
| "loss": 0.3351, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 0.0001428794992175274, | |
| "loss": 0.2785, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 0.00014248826291079813, | |
| "loss": 0.2454, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 0.00014209702660406887, | |
| "loss": 0.4382, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 0.0001417057902973396, | |
| "loss": 0.4117, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 0.00014131455399061033, | |
| "loss": 0.3376, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "eval_accuracy": 0.8315343099460293, | |
| "eval_loss": 0.62732994556427, | |
| "eval_runtime": 43.1453, | |
| "eval_samples_per_second": 60.122, | |
| "eval_steps_per_second": 7.533, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 0.00014092331768388107, | |
| "loss": 0.4258, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 0.0001405320813771518, | |
| "loss": 0.3927, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 0.00014014084507042254, | |
| "loss": 0.4155, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 0.00013974960876369328, | |
| "loss": 0.2391, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 0.00013935837245696402, | |
| "loss": 0.2629, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 0.00013896713615023476, | |
| "loss": 0.3786, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 0.00013857589984350547, | |
| "loss": 0.3195, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 0.00013818466353677621, | |
| "loss": 0.2218, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 0.00013779342723004695, | |
| "loss": 0.2454, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 0.0001374021909233177, | |
| "loss": 0.3071, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "eval_accuracy": 0.8319198149575945, | |
| "eval_loss": 0.5988024473190308, | |
| "eval_runtime": 43.1472, | |
| "eval_samples_per_second": 60.12, | |
| "eval_steps_per_second": 7.532, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 0.0001370109546165884, | |
| "loss": 0.2436, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 0.00013661971830985914, | |
| "loss": 0.244, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 0.0001362284820031299, | |
| "loss": 0.3613, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 0.00013583724569640065, | |
| "loss": 0.4596, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 0.00013544600938967136, | |
| "loss": 0.2966, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 0.0001350547730829421, | |
| "loss": 0.3579, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 0.00013466353677621284, | |
| "loss": 0.3748, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 0.00013427230046948358, | |
| "loss": 0.2644, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 0.0001338810641627543, | |
| "loss": 0.2401, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 0.00013348982785602503, | |
| "loss": 0.2863, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "eval_accuracy": 0.8153430994602929, | |
| "eval_loss": 0.6730513572692871, | |
| "eval_runtime": 42.369, | |
| "eval_samples_per_second": 61.224, | |
| "eval_steps_per_second": 7.671, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 0.00013309859154929577, | |
| "loss": 0.1875, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 0.0001327073552425665, | |
| "loss": 0.2504, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 0.00013231611893583725, | |
| "loss": 0.3071, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 0.000131924882629108, | |
| "loss": 0.3549, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 0.00013153364632237873, | |
| "loss": 0.3535, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 0.00013114241001564947, | |
| "loss": 0.3758, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 0.0001307511737089202, | |
| "loss": 0.3174, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 0.00013035993740219092, | |
| "loss": 0.2612, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 0.00012996870109546166, | |
| "loss": 0.3562, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 0.0001295774647887324, | |
| "loss": 0.3017, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "eval_accuracy": 0.8315343099460293, | |
| "eval_loss": 0.6042011976242065, | |
| "eval_runtime": 43.2556, | |
| "eval_samples_per_second": 59.969, | |
| "eval_steps_per_second": 7.513, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 0.00012918622848200314, | |
| "loss": 0.4068, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 0.00012879499217527385, | |
| "loss": 0.2903, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 0.00012840375586854462, | |
| "loss": 0.2385, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 0.00012801251956181536, | |
| "loss": 0.2733, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 0.0001276212832550861, | |
| "loss": 0.2174, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 0.0001272300469483568, | |
| "loss": 0.2813, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 0.00012683881064162755, | |
| "loss": 0.1874, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 0.0001264475743348983, | |
| "loss": 0.3027, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 0.00012605633802816903, | |
| "loss": 0.3872, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 0.00012566510172143974, | |
| "loss": 0.2382, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "eval_accuracy": 0.8712413261372398, | |
| "eval_loss": 0.5117685794830322, | |
| "eval_runtime": 43.2378, | |
| "eval_samples_per_second": 59.994, | |
| "eval_steps_per_second": 7.517, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 0.00012527386541471048, | |
| "loss": 0.2799, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 0.00012488262910798122, | |
| "loss": 0.2348, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 0.00012449139280125196, | |
| "loss": 0.1834, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 0.0001241001564945227, | |
| "loss": 0.2012, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 0.00012370892018779344, | |
| "loss": 0.182, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 0.00012331768388106418, | |
| "loss": 0.1467, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 0.00012292644757433492, | |
| "loss": 0.2158, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 0.00012253521126760563, | |
| "loss": 0.1666, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 0.00012214397496087637, | |
| "loss": 0.1223, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 0.00012175273865414711, | |
| "loss": 0.1578, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "eval_accuracy": 0.8735543562066307, | |
| "eval_loss": 0.49165645241737366, | |
| "eval_runtime": 42.6596, | |
| "eval_samples_per_second": 60.807, | |
| "eval_steps_per_second": 7.618, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 0.00012136150234741786, | |
| "loss": 0.2114, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 0.00012097026604068857, | |
| "loss": 0.1657, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 0.00012057902973395931, | |
| "loss": 0.1228, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 0.00012018779342723005, | |
| "loss": 0.1246, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 0.00011979655712050079, | |
| "loss": 0.1231, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 0.00011940532081377152, | |
| "loss": 0.2011, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 0.00011901408450704226, | |
| "loss": 0.1177, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 0.000118622848200313, | |
| "loss": 0.1258, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 0.00011823161189358373, | |
| "loss": 0.1668, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 0.00011784037558685446, | |
| "loss": 0.1794, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "eval_accuracy": 0.8631457208943716, | |
| "eval_loss": 0.53019118309021, | |
| "eval_runtime": 43.3542, | |
| "eval_samples_per_second": 59.833, | |
| "eval_steps_per_second": 7.496, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 0.0001174491392801252, | |
| "loss": 0.1714, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 0.00011705790297339594, | |
| "loss": 0.0935, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 0.00011666666666666668, | |
| "loss": 0.1422, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 0.0001162754303599374, | |
| "loss": 0.1965, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 0.00011588419405320814, | |
| "loss": 0.1014, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 0.00011549295774647888, | |
| "loss": 0.2306, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 0.00011510172143974962, | |
| "loss": 0.1482, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 0.00011471048513302034, | |
| "loss": 0.1835, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 0.00011431924882629107, | |
| "loss": 0.0886, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 0.00011392801251956183, | |
| "loss": 0.1093, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "eval_accuracy": 0.8635312259059368, | |
| "eval_loss": 0.5034898519515991, | |
| "eval_runtime": 42.4689, | |
| "eval_samples_per_second": 61.08, | |
| "eval_steps_per_second": 7.653, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 0.00011353677621283257, | |
| "loss": 0.12, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 0.00011314553990610328, | |
| "loss": 0.1595, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 0.00011275430359937402, | |
| "loss": 0.1279, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 0.00011236306729264476, | |
| "loss": 0.1693, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 0.0001119718309859155, | |
| "loss": 0.1546, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 0.00011158059467918622, | |
| "loss": 0.0954, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 0.00011118935837245696, | |
| "loss": 0.1307, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 0.0001107981220657277, | |
| "loss": 0.1041, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 0.00011040688575899844, | |
| "loss": 0.0915, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 0.00011001564945226917, | |
| "loss": 0.1076, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "eval_accuracy": 0.8673862760215882, | |
| "eval_loss": 0.5186420679092407, | |
| "eval_runtime": 42.5346, | |
| "eval_samples_per_second": 60.986, | |
| "eval_steps_per_second": 7.641, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 0.00010962441314553991, | |
| "loss": 0.1075, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 0.00010923317683881065, | |
| "loss": 0.0914, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 0.00010884194053208139, | |
| "loss": 0.1057, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 0.00010845070422535213, | |
| "loss": 0.1967, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 0.00010805946791862285, | |
| "loss": 0.1148, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 0.00010766823161189359, | |
| "loss": 0.0591, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 0.00010727699530516433, | |
| "loss": 0.1369, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 0.00010688575899843507, | |
| "loss": 0.0991, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 0.0001064945226917058, | |
| "loss": 0.1579, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 0.00010610328638497653, | |
| "loss": 0.1219, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "eval_accuracy": 0.8801079414032382, | |
| "eval_loss": 0.4722863733768463, | |
| "eval_runtime": 42.9174, | |
| "eval_samples_per_second": 60.442, | |
| "eval_steps_per_second": 7.573, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 0.00010571205007824727, | |
| "loss": 0.0743, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 0.00010532081377151801, | |
| "loss": 0.1099, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 0.00010492957746478873, | |
| "loss": 0.1275, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 0.00010453834115805947, | |
| "loss": 0.0914, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 0.00010414710485133022, | |
| "loss": 0.1177, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 0.00010375586854460096, | |
| "loss": 0.1255, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 0.00010336463223787167, | |
| "loss": 0.1376, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 0.00010297339593114241, | |
| "loss": 0.1848, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 0.00010258215962441315, | |
| "loss": 0.0986, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 0.0001021909233176839, | |
| "loss": 0.1017, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "eval_accuracy": 0.8712413261372398, | |
| "eval_loss": 0.5132078528404236, | |
| "eval_runtime": 42.9886, | |
| "eval_samples_per_second": 60.342, | |
| "eval_steps_per_second": 7.56, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 0.00010179968701095461, | |
| "loss": 0.1514, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 0.00010140845070422535, | |
| "loss": 0.0884, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 0.00010101721439749609, | |
| "loss": 0.0894, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 0.00010062597809076683, | |
| "loss": 0.0964, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 0.00010023474178403756, | |
| "loss": 0.0851, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 9.98435054773083e-05, | |
| "loss": 0.078, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 9.945226917057904e-05, | |
| "loss": 0.0611, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 9.906103286384976e-05, | |
| "loss": 0.0721, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 9.86697965571205e-05, | |
| "loss": 0.0604, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 9.827856025039124e-05, | |
| "loss": 0.0351, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "eval_accuracy": 0.8727833461835004, | |
| "eval_loss": 0.4709137976169586, | |
| "eval_runtime": 43.1107, | |
| "eval_samples_per_second": 60.171, | |
| "eval_steps_per_second": 7.539, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 9.788732394366198e-05, | |
| "loss": 0.0262, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 9.749608763693271e-05, | |
| "loss": 0.0263, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 9.710485133020345e-05, | |
| "loss": 0.0201, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 9.671361502347419e-05, | |
| "loss": 0.0467, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 9.632237871674493e-05, | |
| "loss": 0.0635, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 9.593114241001565e-05, | |
| "loss": 0.0162, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 9.553990610328639e-05, | |
| "loss": 0.1138, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 9.514866979655712e-05, | |
| "loss": 0.0597, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 9.475743348982786e-05, | |
| "loss": 0.0349, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 9.43661971830986e-05, | |
| "loss": 0.0295, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "eval_accuracy": 0.8824209714726291, | |
| "eval_loss": 0.4673934876918793, | |
| "eval_runtime": 43.0785, | |
| "eval_samples_per_second": 60.216, | |
| "eval_steps_per_second": 7.544, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 9.397496087636933e-05, | |
| "loss": 0.0375, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 9.358372456964006e-05, | |
| "loss": 0.0238, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 9.31924882629108e-05, | |
| "loss": 0.0262, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 9.280125195618154e-05, | |
| "loss": 0.0507, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 9.241001564945228e-05, | |
| "loss": 0.0397, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 9.2018779342723e-05, | |
| "loss": 0.0486, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 9.162754303599374e-05, | |
| "loss": 0.0332, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 9.123630672926447e-05, | |
| "loss": 0.022, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 9.084507042253522e-05, | |
| "loss": 0.0735, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 9.045383411580595e-05, | |
| "loss": 0.0416, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "eval_accuracy": 0.8804934464148034, | |
| "eval_loss": 0.4836331307888031, | |
| "eval_runtime": 42.924, | |
| "eval_samples_per_second": 60.432, | |
| "eval_steps_per_second": 7.572, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 9.006259780907669e-05, | |
| "loss": 0.0188, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 8.967136150234741e-05, | |
| "loss": 0.0361, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 8.928012519561815e-05, | |
| "loss": 0.0655, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 8.888888888888889e-05, | |
| "loss": 0.097, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 8.849765258215963e-05, | |
| "loss": 0.1409, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 8.810641627543036e-05, | |
| "loss": 0.0353, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 8.77151799687011e-05, | |
| "loss": 0.036, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 8.732394366197182e-05, | |
| "loss": 0.0322, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 8.693270735524258e-05, | |
| "loss": 0.1002, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 8.65414710485133e-05, | |
| "loss": 0.0386, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "eval_accuracy": 0.8828064764841943, | |
| "eval_loss": 0.466279000043869, | |
| "eval_runtime": 43.3083, | |
| "eval_samples_per_second": 59.896, | |
| "eval_steps_per_second": 7.504, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 8.615023474178404e-05, | |
| "loss": 0.0219, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 8.575899843505478e-05, | |
| "loss": 0.0428, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 8.536776212832551e-05, | |
| "loss": 0.0941, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 8.497652582159625e-05, | |
| "loss": 0.061, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 8.458528951486699e-05, | |
| "loss": 0.0524, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 8.419405320813773e-05, | |
| "loss": 0.0961, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 8.380281690140845e-05, | |
| "loss": 0.0524, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 8.341158059467919e-05, | |
| "loss": 0.0604, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 8.302034428794993e-05, | |
| "loss": 0.0534, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 8.262910798122067e-05, | |
| "loss": 0.0392, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "eval_accuracy": 0.8989976869699307, | |
| "eval_loss": 0.4003235101699829, | |
| "eval_runtime": 43.2036, | |
| "eval_samples_per_second": 60.041, | |
| "eval_steps_per_second": 7.523, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 8.22378716744914e-05, | |
| "loss": 0.0273, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 8.184663536776213e-05, | |
| "loss": 0.0529, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 8.145539906103286e-05, | |
| "loss": 0.0726, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 8.106416275430361e-05, | |
| "loss": 0.0505, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 8.067292644757434e-05, | |
| "loss": 0.0778, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 8.028169014084508e-05, | |
| "loss": 0.0891, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 7.98904538341158e-05, | |
| "loss": 0.0264, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 7.949921752738654e-05, | |
| "loss": 0.0112, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 7.910798122065728e-05, | |
| "loss": 0.0163, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 7.871674491392802e-05, | |
| "loss": 0.0383, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "eval_accuracy": 0.894757131842714, | |
| "eval_loss": 0.4187348186969757, | |
| "eval_runtime": 43.069, | |
| "eval_samples_per_second": 60.229, | |
| "eval_steps_per_second": 7.546, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 7.832550860719875e-05, | |
| "loss": 0.0867, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 7.793427230046949e-05, | |
| "loss": 0.0649, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 7.754303599374021e-05, | |
| "loss": 0.0959, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 7.715179968701097e-05, | |
| "loss": 0.0634, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 7.676056338028169e-05, | |
| "loss": 0.0607, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 7.636932707355243e-05, | |
| "loss": 0.0246, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 7.597809076682316e-05, | |
| "loss": 0.0421, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 7.55868544600939e-05, | |
| "loss": 0.0324, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 7.519561815336464e-05, | |
| "loss": 0.0119, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 5.01, | |
| "learning_rate": 7.480438184663538e-05, | |
| "loss": 0.0624, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 5.01, | |
| "eval_accuracy": 0.887432536622976, | |
| "eval_loss": 0.44595664739608765, | |
| "eval_runtime": 42.8626, | |
| "eval_samples_per_second": 60.519, | |
| "eval_steps_per_second": 7.582, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 5.02, | |
| "learning_rate": 7.44131455399061e-05, | |
| "loss": 0.0158, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 7.402190923317684e-05, | |
| "loss": 0.0264, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 5.05, | |
| "learning_rate": 7.363067292644758e-05, | |
| "loss": 0.0129, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 5.07, | |
| "learning_rate": 7.323943661971832e-05, | |
| "loss": 0.0153, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 5.09, | |
| "learning_rate": 7.284820031298905e-05, | |
| "loss": 0.0111, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 5.1, | |
| "learning_rate": 7.245696400625979e-05, | |
| "loss": 0.0179, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 5.12, | |
| "learning_rate": 7.206572769953051e-05, | |
| "loss": 0.0174, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 5.13, | |
| "learning_rate": 7.167449139280125e-05, | |
| "loss": 0.0396, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 5.15, | |
| "learning_rate": 7.128325508607199e-05, | |
| "loss": 0.0073, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 7.089201877934273e-05, | |
| "loss": 0.0188, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "eval_accuracy": 0.9028527370855821, | |
| "eval_loss": 0.41694527864456177, | |
| "eval_runtime": 42.4867, | |
| "eval_samples_per_second": 61.054, | |
| "eval_steps_per_second": 7.649, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 5.18, | |
| "learning_rate": 7.050078247261346e-05, | |
| "loss": 0.0048, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 7.01095461658842e-05, | |
| "loss": 0.006, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 5.21, | |
| "learning_rate": 6.971830985915493e-05, | |
| "loss": 0.018, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 5.23, | |
| "learning_rate": 6.932707355242567e-05, | |
| "loss": 0.0314, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 6.89358372456964e-05, | |
| "loss": 0.0113, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 6.854460093896714e-05, | |
| "loss": 0.0075, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 5.27, | |
| "learning_rate": 6.815336463223787e-05, | |
| "loss": 0.0148, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 6.77621283255086e-05, | |
| "loss": 0.0092, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 6.737089201877934e-05, | |
| "loss": 0.022, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 5.32, | |
| "learning_rate": 6.697965571205008e-05, | |
| "loss": 0.0174, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 5.32, | |
| "eval_accuracy": 0.8951426368542791, | |
| "eval_loss": 0.40977469086647034, | |
| "eval_runtime": 42.8425, | |
| "eval_samples_per_second": 60.547, | |
| "eval_steps_per_second": 7.586, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 5.34, | |
| "learning_rate": 6.658841940532081e-05, | |
| "loss": 0.0054, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 5.35, | |
| "learning_rate": 6.619718309859155e-05, | |
| "loss": 0.0051, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 5.37, | |
| "learning_rate": 6.580594679186229e-05, | |
| "loss": 0.0316, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 5.38, | |
| "learning_rate": 6.541471048513303e-05, | |
| "loss": 0.0127, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 5.4, | |
| "learning_rate": 6.502347417840375e-05, | |
| "loss": 0.0085, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 6.463223787167449e-05, | |
| "loss": 0.0383, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 5.43, | |
| "learning_rate": 6.424100156494522e-05, | |
| "loss": 0.0075, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 5.45, | |
| "learning_rate": 6.384976525821597e-05, | |
| "loss": 0.0049, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 5.46, | |
| "learning_rate": 6.345852895148671e-05, | |
| "loss": 0.0259, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 5.48, | |
| "learning_rate": 6.306729264475744e-05, | |
| "loss": 0.0257, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 5.48, | |
| "eval_accuracy": 0.8951426368542791, | |
| "eval_loss": 0.4288833737373352, | |
| "eval_runtime": 42.534, | |
| "eval_samples_per_second": 60.987, | |
| "eval_steps_per_second": 7.641, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 5.49, | |
| "learning_rate": 6.267605633802818e-05, | |
| "loss": 0.0194, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 6.22848200312989e-05, | |
| "loss": 0.0144, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 5.52, | |
| "learning_rate": 6.189358372456964e-05, | |
| "loss": 0.0128, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 5.54, | |
| "learning_rate": 6.150234741784038e-05, | |
| "loss": 0.0273, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 5.56, | |
| "learning_rate": 6.111111111111112e-05, | |
| "loss": 0.0447, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 6.071987480438185e-05, | |
| "loss": 0.0389, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 5.59, | |
| "learning_rate": 6.032863849765259e-05, | |
| "loss": 0.0288, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 5.993740219092332e-05, | |
| "loss": 0.0101, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 5.62, | |
| "learning_rate": 5.954616588419406e-05, | |
| "loss": 0.006, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "learning_rate": 5.915492957746479e-05, | |
| "loss": 0.0123, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "eval_accuracy": 0.9028527370855821, | |
| "eval_loss": 0.4294571876525879, | |
| "eval_runtime": 42.3329, | |
| "eval_samples_per_second": 61.276, | |
| "eval_steps_per_second": 7.677, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 5.65, | |
| "learning_rate": 5.876369327073553e-05, | |
| "loss": 0.017, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 5.67, | |
| "learning_rate": 5.837245696400626e-05, | |
| "loss": 0.0132, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 5.68, | |
| "learning_rate": 5.7981220657277e-05, | |
| "loss": 0.0201, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 5.7, | |
| "learning_rate": 5.7589984350547735e-05, | |
| "loss": 0.0104, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 5.7198748043818474e-05, | |
| "loss": 0.0054, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 5.73, | |
| "learning_rate": 5.68075117370892e-05, | |
| "loss": 0.0137, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 5.6416275430359946e-05, | |
| "loss": 0.0038, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 5.76, | |
| "learning_rate": 5.602503912363067e-05, | |
| "loss": 0.0052, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 5.563380281690141e-05, | |
| "loss": 0.0142, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 5.79, | |
| "learning_rate": 5.5242566510172144e-05, | |
| "loss": 0.0052, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 5.79, | |
| "eval_accuracy": 0.8993831919814957, | |
| "eval_loss": 0.43951472640037537, | |
| "eval_runtime": 42.973, | |
| "eval_samples_per_second": 60.364, | |
| "eval_steps_per_second": 7.563, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 5.81, | |
| "learning_rate": 5.4851330203442884e-05, | |
| "loss": 0.0176, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 5.4460093896713616e-05, | |
| "loss": 0.0226, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 5.84, | |
| "learning_rate": 5.4068857589984356e-05, | |
| "loss": 0.007, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 5.85, | |
| "learning_rate": 5.367762128325509e-05, | |
| "loss": 0.0444, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 5.87, | |
| "learning_rate": 5.328638497652583e-05, | |
| "loss": 0.0056, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 5.2895148669796554e-05, | |
| "loss": 0.0172, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 5.9, | |
| "learning_rate": 5.25039123630673e-05, | |
| "loss": 0.0235, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 5.92, | |
| "learning_rate": 5.2112676056338026e-05, | |
| "loss": 0.0065, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 5.172143974960877e-05, | |
| "loss": 0.0083, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 5.95, | |
| "learning_rate": 5.13302034428795e-05, | |
| "loss": 0.0081, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 5.95, | |
| "eval_accuracy": 0.9082498072474943, | |
| "eval_loss": 0.42170995473861694, | |
| "eval_runtime": 43.0345, | |
| "eval_samples_per_second": 60.277, | |
| "eval_steps_per_second": 7.552, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 5.093896713615024e-05, | |
| "loss": 0.0071, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 5.98, | |
| "learning_rate": 5.054773082942097e-05, | |
| "loss": 0.031, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 5.99, | |
| "learning_rate": 5.015649452269171e-05, | |
| "loss": 0.0057, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 6.01, | |
| "learning_rate": 4.976525821596245e-05, | |
| "loss": 0.0051, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 6.03, | |
| "learning_rate": 4.937402190923318e-05, | |
| "loss": 0.0064, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 6.04, | |
| "learning_rate": 4.8982785602503914e-05, | |
| "loss": 0.0033, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 6.06, | |
| "learning_rate": 4.8591549295774653e-05, | |
| "loss": 0.0033, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 6.07, | |
| "learning_rate": 4.8200312989045386e-05, | |
| "loss": 0.0032, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 4.7809076682316126e-05, | |
| "loss": 0.0033, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 6.1, | |
| "learning_rate": 4.741784037558686e-05, | |
| "loss": 0.0032, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 6.1, | |
| "eval_accuracy": 0.9055512721665382, | |
| "eval_loss": 0.42163705825805664, | |
| "eval_runtime": 43.2354, | |
| "eval_samples_per_second": 59.997, | |
| "eval_steps_per_second": 7.517, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 6.12, | |
| "learning_rate": 4.702660406885759e-05, | |
| "loss": 0.0035, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 6.13, | |
| "learning_rate": 4.663536776212833e-05, | |
| "loss": 0.0032, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 6.15, | |
| "learning_rate": 4.624413145539906e-05, | |
| "loss": 0.003, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 6.17, | |
| "learning_rate": 4.58528951486698e-05, | |
| "loss": 0.0031, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 6.18, | |
| "learning_rate": 4.5461658841940535e-05, | |
| "loss": 0.0029, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 6.2, | |
| "learning_rate": 4.507042253521127e-05, | |
| "loss": 0.0115, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 4.467918622848201e-05, | |
| "loss": 0.0037, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 4.428794992175274e-05, | |
| "loss": 0.0056, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 6.24, | |
| "learning_rate": 4.389671361502348e-05, | |
| "loss": 0.0027, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 6.26, | |
| "learning_rate": 4.350547730829421e-05, | |
| "loss": 0.0033, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 6.26, | |
| "eval_accuracy": 0.9082498072474943, | |
| "eval_loss": 0.4112599790096283, | |
| "eval_runtime": 42.7966, | |
| "eval_samples_per_second": 60.612, | |
| "eval_steps_per_second": 7.594, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 6.28, | |
| "learning_rate": 4.311424100156495e-05, | |
| "loss": 0.0044, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 4.2723004694835684e-05, | |
| "loss": 0.0033, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 4.2331768388106416e-05, | |
| "loss": 0.0031, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 6.32, | |
| "learning_rate": 4.1940532081377156e-05, | |
| "loss": 0.0033, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 6.34, | |
| "learning_rate": 4.154929577464789e-05, | |
| "loss": 0.0048, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 6.35, | |
| "learning_rate": 4.115805946791863e-05, | |
| "loss": 0.003, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 6.37, | |
| "learning_rate": 4.076682316118936e-05, | |
| "loss": 0.0044, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 6.38, | |
| "learning_rate": 4.037558685446009e-05, | |
| "loss": 0.0037, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 3.998435054773083e-05, | |
| "loss": 0.0024, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "learning_rate": 3.9593114241001565e-05, | |
| "loss": 0.0024, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "eval_accuracy": 0.9101773323053199, | |
| "eval_loss": 0.40595710277557373, | |
| "eval_runtime": 42.4804, | |
| "eval_samples_per_second": 61.063, | |
| "eval_steps_per_second": 7.651, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "learning_rate": 3.9201877934272305e-05, | |
| "loss": 0.0027, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 6.45, | |
| "learning_rate": 3.881064162754304e-05, | |
| "loss": 0.0036, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 6.46, | |
| "learning_rate": 3.841940532081377e-05, | |
| "loss": 0.0029, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 6.48, | |
| "learning_rate": 3.802816901408451e-05, | |
| "loss": 0.0057, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 3.763693270735524e-05, | |
| "loss": 0.0033, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 6.51, | |
| "learning_rate": 3.724569640062598e-05, | |
| "loss": 0.003, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 3.6854460093896714e-05, | |
| "loss": 0.0028, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 6.54, | |
| "learning_rate": 3.646322378716745e-05, | |
| "loss": 0.0028, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 6.56, | |
| "learning_rate": 3.6071987480438186e-05, | |
| "loss": 0.0075, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 3.568075117370892e-05, | |
| "loss": 0.0022, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "eval_accuracy": 0.9090208172706246, | |
| "eval_loss": 0.4067469835281372, | |
| "eval_runtime": 42.6487, | |
| "eval_samples_per_second": 60.823, | |
| "eval_steps_per_second": 7.62, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 6.59, | |
| "learning_rate": 3.528951486697966e-05, | |
| "loss": 0.0046, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 3.489827856025039e-05, | |
| "loss": 0.0055, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 6.62, | |
| "learning_rate": 3.450704225352113e-05, | |
| "loss": 0.0026, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 6.64, | |
| "learning_rate": 3.411580594679186e-05, | |
| "loss": 0.0028, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 6.65, | |
| "learning_rate": 3.3724569640062596e-05, | |
| "loss": 0.005, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 6.67, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 0.0028, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 6.68, | |
| "learning_rate": 3.294209702660407e-05, | |
| "loss": 0.0023, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 6.7, | |
| "learning_rate": 3.255086071987481e-05, | |
| "loss": 0.0026, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 6.71, | |
| "learning_rate": 3.215962441314554e-05, | |
| "loss": 0.0025, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 6.73, | |
| "learning_rate": 3.176838810641627e-05, | |
| "loss": 0.0031, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 6.73, | |
| "eval_accuracy": 0.9113338473400154, | |
| "eval_loss": 0.4005250632762909, | |
| "eval_runtime": 43.1239, | |
| "eval_samples_per_second": 60.152, | |
| "eval_steps_per_second": 7.536, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 6.74, | |
| "learning_rate": 3.137715179968701e-05, | |
| "loss": 0.0027, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 6.76, | |
| "learning_rate": 3.0985915492957744e-05, | |
| "loss": 0.0024, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 6.78, | |
| "learning_rate": 3.0594679186228484e-05, | |
| "loss": 0.0031, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 6.79, | |
| "learning_rate": 3.0203442879499216e-05, | |
| "loss": 0.0109, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 6.81, | |
| "learning_rate": 2.9812206572769952e-05, | |
| "loss": 0.0032, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 2.9420970266040692e-05, | |
| "loss": 0.0027, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 6.84, | |
| "learning_rate": 2.9029733959311428e-05, | |
| "loss": 0.0033, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 2.8638497652582164e-05, | |
| "loss": 0.0044, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 2.82472613458529e-05, | |
| "loss": 0.0029, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 2.7856025039123636e-05, | |
| "loss": 0.0021, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "eval_accuracy": 0.912875867386276, | |
| "eval_loss": 0.40083467960357666, | |
| "eval_runtime": 43.0173, | |
| "eval_samples_per_second": 60.301, | |
| "eval_steps_per_second": 7.555, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 6.9, | |
| "learning_rate": 2.746478873239437e-05, | |
| "loss": 0.0021, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 6.92, | |
| "learning_rate": 2.7073552425665105e-05, | |
| "loss": 0.0031, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 6.93, | |
| "learning_rate": 2.668231611893584e-05, | |
| "loss": 0.0023, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 6.95, | |
| "learning_rate": 2.6291079812206577e-05, | |
| "loss": 0.0032, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 6.96, | |
| "learning_rate": 2.5899843505477313e-05, | |
| "loss": 0.0339, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 2.5508607198748045e-05, | |
| "loss": 0.0025, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "learning_rate": 2.511737089201878e-05, | |
| "loss": 0.0057, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 7.01, | |
| "learning_rate": 2.4726134585289514e-05, | |
| "loss": 0.0029, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 7.03, | |
| "learning_rate": 2.433489827856025e-05, | |
| "loss": 0.0019, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 2.3943661971830986e-05, | |
| "loss": 0.0021, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "eval_accuracy": 0.9113338473400154, | |
| "eval_loss": 0.3966985046863556, | |
| "eval_runtime": 43.1598, | |
| "eval_samples_per_second": 60.102, | |
| "eval_steps_per_second": 7.53, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 2.3552425665101726e-05, | |
| "loss": 0.0034, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 7.07, | |
| "learning_rate": 2.3161189358372458e-05, | |
| "loss": 0.0023, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 2.2769953051643194e-05, | |
| "loss": 0.0025, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 7.1, | |
| "learning_rate": 2.237871674491393e-05, | |
| "loss": 0.0021, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 7.12, | |
| "learning_rate": 2.1987480438184666e-05, | |
| "loss": 0.0025, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 2.1596244131455402e-05, | |
| "loss": 0.0025, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 7.15, | |
| "learning_rate": 2.1205007824726135e-05, | |
| "loss": 0.0025, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 7.17, | |
| "learning_rate": 2.081377151799687e-05, | |
| "loss": 0.0023, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 7.18, | |
| "learning_rate": 2.0422535211267607e-05, | |
| "loss": 0.0023, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "learning_rate": 2.0031298904538343e-05, | |
| "loss": 0.0043, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "eval_accuracy": 0.9121048573631457, | |
| "eval_loss": 0.3959566652774811, | |
| "eval_runtime": 43.6577, | |
| "eval_samples_per_second": 59.417, | |
| "eval_steps_per_second": 7.444, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 7.21, | |
| "learning_rate": 1.964006259780908e-05, | |
| "loss": 0.0021, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 7.23, | |
| "learning_rate": 1.9248826291079812e-05, | |
| "loss": 0.0024, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 7.25, | |
| "learning_rate": 1.8857589984350548e-05, | |
| "loss": 0.0023, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 7.26, | |
| "learning_rate": 1.8466353677621284e-05, | |
| "loss": 0.0019, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 7.28, | |
| "learning_rate": 1.807511737089202e-05, | |
| "loss": 0.002, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 7.29, | |
| "learning_rate": 1.7683881064162756e-05, | |
| "loss": 0.0023, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 1.7292644757433492e-05, | |
| "loss": 0.0022, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 7.32, | |
| "learning_rate": 1.6901408450704224e-05, | |
| "loss": 0.0021, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 1.651017214397496e-05, | |
| "loss": 0.0021, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 1.6118935837245697e-05, | |
| "loss": 0.0022, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "eval_accuracy": 0.9124903623747108, | |
| "eval_loss": 0.3962063193321228, | |
| "eval_runtime": 43.4452, | |
| "eval_samples_per_second": 59.707, | |
| "eval_steps_per_second": 7.481, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 7.37, | |
| "learning_rate": 1.5727699530516433e-05, | |
| "loss": 0.0023, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 7.39, | |
| "learning_rate": 1.533646322378717e-05, | |
| "loss": 0.0028, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 7.4, | |
| "learning_rate": 1.4945226917057903e-05, | |
| "loss": 0.0022, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 7.42, | |
| "learning_rate": 1.4553990610328639e-05, | |
| "loss": 0.0023, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 1.4162754303599373e-05, | |
| "loss": 0.0025, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 7.45, | |
| "learning_rate": 1.377151799687011e-05, | |
| "loss": 0.013, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 7.46, | |
| "learning_rate": 1.3380281690140845e-05, | |
| "loss": 0.0021, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 7.48, | |
| "learning_rate": 1.298904538341158e-05, | |
| "loss": 0.0023, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "learning_rate": 1.2597809076682316e-05, | |
| "loss": 0.0027, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 7.51, | |
| "learning_rate": 1.2206572769953052e-05, | |
| "loss": 0.0021, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 7.51, | |
| "eval_accuracy": 0.9121048573631457, | |
| "eval_loss": 0.39916661381721497, | |
| "eval_runtime": 43.2453, | |
| "eval_samples_per_second": 59.983, | |
| "eval_steps_per_second": 7.515, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 7.53, | |
| "learning_rate": 1.1815336463223788e-05, | |
| "loss": 0.0021, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 7.54, | |
| "learning_rate": 1.1424100156494522e-05, | |
| "loss": 0.002, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 7.56, | |
| "learning_rate": 1.1032863849765258e-05, | |
| "loss": 0.0019, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 7.57, | |
| "learning_rate": 1.0641627543035994e-05, | |
| "loss": 0.0025, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 7.59, | |
| "learning_rate": 1.0250391236306729e-05, | |
| "loss": 0.0023, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 7.61, | |
| "learning_rate": 9.859154929577465e-06, | |
| "loss": 0.0022, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 9.4679186228482e-06, | |
| "loss": 0.0108, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 7.64, | |
| "learning_rate": 9.076682316118937e-06, | |
| "loss": 0.0021, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 7.65, | |
| "learning_rate": 8.685446009389673e-06, | |
| "loss": 0.0023, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "learning_rate": 8.294209702660407e-06, | |
| "loss": 0.002, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "eval_accuracy": 0.912875867386276, | |
| "eval_loss": 0.39511793851852417, | |
| "eval_runtime": 43.4397, | |
| "eval_samples_per_second": 59.715, | |
| "eval_steps_per_second": 7.482, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 7.68, | |
| "learning_rate": 7.902973395931143e-06, | |
| "loss": 0.0024, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 7.7, | |
| "learning_rate": 7.511737089201878e-06, | |
| "loss": 0.0021, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 7.72, | |
| "learning_rate": 7.120500782472614e-06, | |
| "loss": 0.0019, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 7.73, | |
| "learning_rate": 6.7292644757433494e-06, | |
| "loss": 0.0022, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 7.75, | |
| "learning_rate": 6.338028169014085e-06, | |
| "loss": 0.0023, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 7.76, | |
| "learning_rate": 5.94679186228482e-06, | |
| "loss": 0.002, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 5.555555555555556e-06, | |
| "loss": 0.0017, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 7.79, | |
| "learning_rate": 5.164319248826292e-06, | |
| "loss": 0.0024, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 7.81, | |
| "learning_rate": 4.773082942097027e-06, | |
| "loss": 0.0024, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 7.82, | |
| "learning_rate": 4.381846635367762e-06, | |
| "loss": 0.0023, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 7.82, | |
| "eval_accuracy": 0.9124903623747108, | |
| "eval_loss": 0.3951539695262909, | |
| "eval_runtime": 42.332, | |
| "eval_samples_per_second": 61.278, | |
| "eval_steps_per_second": 7.677, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 7.84, | |
| "learning_rate": 3.990610328638498e-06, | |
| "loss": 0.0021, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 7.86, | |
| "learning_rate": 3.5993740219092334e-06, | |
| "loss": 0.0025, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 3.2081377151799686e-06, | |
| "loss": 0.0029, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 2.8169014084507042e-06, | |
| "loss": 0.0024, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 7.9, | |
| "learning_rate": 2.42566510172144e-06, | |
| "loss": 0.0019, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 7.92, | |
| "learning_rate": 2.0344287949921754e-06, | |
| "loss": 0.0022, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 7.93, | |
| "learning_rate": 1.643192488262911e-06, | |
| "loss": 0.0025, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 7.95, | |
| "learning_rate": 1.2519561815336464e-06, | |
| "loss": 0.0018, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 7.97, | |
| "learning_rate": 8.607198748043818e-07, | |
| "loss": 0.0017, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 7.98, | |
| "learning_rate": 4.694835680751174e-07, | |
| "loss": 0.0021, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 7.98, | |
| "eval_accuracy": 0.912875867386276, | |
| "eval_loss": 0.39521878957748413, | |
| "eval_runtime": 43.0806, | |
| "eval_samples_per_second": 60.213, | |
| "eval_steps_per_second": 7.544, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 7.82472613458529e-08, | |
| "loss": 0.0026, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "step": 5112, | |
| "total_flos": 6.337884979995771e+18, | |
| "train_loss": 0.4943873079753071, | |
| "train_runtime": 5596.588, | |
| "train_samples_per_second": 14.595, | |
| "train_steps_per_second": 0.913 | |
| } | |
| ], | |
| "max_steps": 5112, | |
| "num_train_epochs": 8, | |
| "total_flos": 6.337884979995771e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |