| { |
| "best_metric": 0.9291033148765564, |
| "best_model_checkpoint": "segformer-finetuned-segments-cmp-facade-outputs-transforms/checkpoint-360", |
| "epoch": 50.0, |
| "global_step": 1600, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.03, |
| "learning_rate": 0.000599625, |
| "loss": 1.4974, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 0.00059925, |
| "loss": 1.5572, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 0.000598875, |
| "loss": 1.5276, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 0.0005985, |
| "loss": 1.5646, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 0.0005981249999999999, |
| "loss": 1.3389, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 0.00059775, |
| "loss": 1.5391, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 0.0005973749999999999, |
| "loss": 1.3854, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 0.000597, |
| "loss": 1.4288, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 0.0005966249999999999, |
| "loss": 1.3674, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 0.00059625, |
| "loss": 1.2815, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 0.0005958749999999999, |
| "loss": 1.3141, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 0.0005955, |
| "loss": 1.2826, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 0.000595125, |
| "loss": 1.2023, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 0.0005947499999999999, |
| "loss": 1.2487, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 0.000594375, |
| "loss": 1.2041, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 0.0005939999999999999, |
| "loss": 1.3763, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 0.000593625, |
| "loss": 1.2913, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 0.0005932499999999999, |
| "loss": 1.2048, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 0.000592875, |
| "loss": 1.2608, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 0.0005924999999999999, |
| "loss": 1.2328, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.62, |
| "eval_loss": 1.2321703433990479, |
| "eval_mean_accuracy": 0.27640981040002394, |
| "eval_mean_iou": 0.19381676955791272, |
| "eval_overall_accuracy": 0.5843182279352557, |
| "eval_per_category_accuracy": [ |
| 0.6484014402727478, |
| 0.7268698138428387, |
| 0.7396621827345261, |
| 0.0, |
| 0.314071721025026, |
| 0.16580774409430243, |
| 0.18421110637031626, |
| 0.0, |
| 0.0, |
| 0.48692030347962645, |
| 0.0, |
| 0.05097341298090315 |
| ], |
| "eval_per_category_iou": [ |
| 0.46588897486611575, |
| 0.4794379279754499, |
| 0.47516306513696516, |
| 0.0, |
| 0.21479697731322273, |
| 0.1468532929608522, |
| 0.15399766900533918, |
| 0.0, |
| 0.0, |
| 0.34140124873016997, |
| 0.0, |
| 0.04826207870683769 |
| ], |
| "eval_runtime": 121.98, |
| "eval_samples_per_second": 0.935, |
| "eval_steps_per_second": 0.082, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 0.0005921249999999999, |
| "loss": 1.2252, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 0.0005917499999999999, |
| "loss": 1.1954, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 0.0005913749999999999, |
| "loss": 1.2127, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 0.0005909999999999999, |
| "loss": 1.1908, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 0.0005906249999999999, |
| "loss": 1.1458, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 0.00059025, |
| "loss": 1.0854, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 0.000589875, |
| "loss": 1.1372, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 0.0005895, |
| "loss": 1.0638, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 0.000589125, |
| "loss": 1.1055, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 0.00058875, |
| "loss": 1.1073, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 0.0005883749999999999, |
| "loss": 1.0367, |
| "step": 31 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 0.000588, |
| "loss": 1.2498, |
| "step": 32 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 0.0005876249999999999, |
| "loss": 1.2009, |
| "step": 33 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 0.00058725, |
| "loss": 1.2267, |
| "step": 34 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 0.0005868749999999999, |
| "loss": 0.9791, |
| "step": 35 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 0.0005865, |
| "loss": 1.0073, |
| "step": 36 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 0.000586125, |
| "loss": 1.0052, |
| "step": 37 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 0.0005857499999999999, |
| "loss": 1.0387, |
| "step": 38 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 0.000585375, |
| "loss": 1.076, |
| "step": 39 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 0.0005849999999999999, |
| "loss": 1.1158, |
| "step": 40 |
| }, |
| { |
| "epoch": 1.25, |
| "eval_loss": 1.1340055465698242, |
| "eval_mean_accuracy": 0.3584914050139854, |
| "eval_mean_iou": 0.25033676612937344, |
| "eval_overall_accuracy": 0.6050458539996231, |
| "eval_per_category_accuracy": [ |
| 0.8000948297521079, |
| 0.6337371473364622, |
| 0.6603131057495936, |
| 0.003965968313890952, |
| 0.34629573729708046, |
| 0.3680114248799625, |
| 0.4865078902745655, |
| 0.04561502007369462, |
| 0.0118031463226356, |
| 0.5511836320859765, |
| 0.0, |
| 0.3943689580818554 |
| ], |
| "eval_per_category_iou": [ |
| 0.5176739126303989, |
| 0.47108775141725884, |
| 0.49329266238923963, |
| 0.003963836494345551, |
| 0.26672845921223864, |
| 0.2701529378697397, |
| 0.31332750965950773, |
| 0.04532147556410988, |
| 0.011623004855842847, |
| 0.3621998398780441, |
| 0.0, |
| 0.24866980358175594 |
| ], |
| "eval_runtime": 124.8448, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 40 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 0.000584625, |
| "loss": 1.0057, |
| "step": 41 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 0.0005842499999999999, |
| "loss": 0.9904, |
| "step": 42 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 0.000583875, |
| "loss": 1.1023, |
| "step": 43 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 0.0005834999999999999, |
| "loss": 1.1263, |
| "step": 44 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 0.000583125, |
| "loss": 0.9288, |
| "step": 45 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 0.0005827499999999999, |
| "loss": 0.9942, |
| "step": 46 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 0.0005823749999999999, |
| "loss": 0.9388, |
| "step": 47 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 0.0005819999999999999, |
| "loss": 1.0192, |
| "step": 48 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 0.0005816249999999999, |
| "loss": 1.0228, |
| "step": 49 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 0.00058125, |
| "loss": 0.9034, |
| "step": 50 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 0.000580875, |
| "loss": 1.0346, |
| "step": 51 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 0.0005805, |
| "loss": 0.8394, |
| "step": 52 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 0.000580125, |
| "loss": 0.9892, |
| "step": 53 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 0.00057975, |
| "loss": 1.0063, |
| "step": 54 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 0.0005793749999999999, |
| "loss": 1.0106, |
| "step": 55 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 0.000579, |
| "loss": 0.9481, |
| "step": 56 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 0.0005786249999999999, |
| "loss": 1.0124, |
| "step": 57 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 0.00057825, |
| "loss": 0.9676, |
| "step": 58 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 0.0005778749999999999, |
| "loss": 0.7901, |
| "step": 59 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 0.0005775, |
| "loss": 0.9771, |
| "step": 60 |
| }, |
| { |
| "epoch": 1.88, |
| "eval_loss": 1.0994304418563843, |
| "eval_mean_accuracy": 0.3909737030282887, |
| "eval_mean_iou": 0.2835007332832486, |
| "eval_overall_accuracy": 0.6179555591784025, |
| "eval_per_category_accuracy": [ |
| 0.7168020611741474, |
| 0.7219130768570321, |
| 0.5340207091634352, |
| 0.0431815185182615, |
| 0.45474233550172505, |
| 0.5430138746239189, |
| 0.469268628306114, |
| 0.217241626853214, |
| 0.09605778018331367, |
| 0.5683827928074056, |
| 0.0005080911993094857, |
| 0.3265519411515883 |
| ], |
| "eval_per_category_iou": [ |
| 0.5113476068813879, |
| 0.49334140108090285, |
| 0.4746839761440892, |
| 0.040214174001835405, |
| 0.31186037529280675, |
| 0.31181019822943934, |
| 0.35578648191496853, |
| 0.20097078621392386, |
| 0.08022661461990363, |
| 0.38374995642386217, |
| 0.0005073985432159898, |
| 0.23750983005264706 |
| ], |
| "eval_runtime": 125.5081, |
| "eval_samples_per_second": 0.908, |
| "eval_steps_per_second": 0.08, |
| "step": 60 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 0.000577125, |
| "loss": 0.9391, |
| "step": 61 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 0.00057675, |
| "loss": 1.0473, |
| "step": 62 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 0.000576375, |
| "loss": 0.918, |
| "step": 63 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 0.0005759999999999999, |
| "loss": 1.0355, |
| "step": 64 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 0.000575625, |
| "loss": 1.015, |
| "step": 65 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 0.0005752499999999999, |
| "loss": 0.9346, |
| "step": 66 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 0.000574875, |
| "loss": 0.8417, |
| "step": 67 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 0.0005744999999999999, |
| "loss": 0.8185, |
| "step": 68 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 0.000574125, |
| "loss": 0.9263, |
| "step": 69 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 0.0005737499999999999, |
| "loss": 0.8448, |
| "step": 70 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 0.0005733749999999999, |
| "loss": 1.1335, |
| "step": 71 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 0.0005729999999999999, |
| "loss": 0.9229, |
| "step": 72 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 0.0005726249999999999, |
| "loss": 0.8549, |
| "step": 73 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 0.00057225, |
| "loss": 0.8818, |
| "step": 74 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 0.0005718749999999999, |
| "loss": 0.831, |
| "step": 75 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 0.0005715, |
| "loss": 0.8845, |
| "step": 76 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 0.000571125, |
| "loss": 0.9197, |
| "step": 77 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 0.00057075, |
| "loss": 0.8579, |
| "step": 78 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 0.000570375, |
| "loss": 0.8167, |
| "step": 79 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 0.00057, |
| "loss": 0.8566, |
| "step": 80 |
| }, |
| { |
| "epoch": 2.5, |
| "eval_loss": 1.0861966609954834, |
| "eval_mean_accuracy": 0.3778088961525232, |
| "eval_mean_iou": 0.26404526643863496, |
| "eval_overall_accuracy": 0.6308857432582922, |
| "eval_per_category_accuracy": [ |
| 0.6129614268456303, |
| 0.7713753471720866, |
| 0.7994049832227069, |
| 0.019437755463182074, |
| 0.5688067632752372, |
| 0.4619781591716599, |
| 0.3827566445038881, |
| 0.08782710157707345, |
| 0.15580996069314926, |
| 0.5327405741537456, |
| 0.0009059939457566733, |
| 0.1397020438061634 |
| ], |
| "eval_per_category_iou": [ |
| 0.5101947204251175, |
| 0.5344538212406682, |
| 0.49436541874132667, |
| 0.018884843581445522, |
| 0.270377214768106, |
| 0.28449876963719384, |
| 0.32160160861987425, |
| 0.086283029811171, |
| 0.11932598089180631, |
| 0.39498832385101396, |
| 0.0009054175167701066, |
| 0.13266404817912572 |
| ], |
| "eval_runtime": 123.9, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 80 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 0.0005696249999999999, |
| "loss": 0.827, |
| "step": 81 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 0.00056925, |
| "loss": 0.9862, |
| "step": 82 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 0.0005688749999999999, |
| "loss": 0.8236, |
| "step": 83 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 0.0005685, |
| "loss": 0.8522, |
| "step": 84 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 0.0005681249999999999, |
| "loss": 0.9125, |
| "step": 85 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 0.00056775, |
| "loss": 0.8568, |
| "step": 86 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 0.000567375, |
| "loss": 0.8791, |
| "step": 87 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 0.0005669999999999999, |
| "loss": 0.8188, |
| "step": 88 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 0.000566625, |
| "loss": 0.992, |
| "step": 89 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 0.0005662499999999999, |
| "loss": 0.8239, |
| "step": 90 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 0.000565875, |
| "loss": 0.8892, |
| "step": 91 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 0.0005654999999999999, |
| "loss": 0.8093, |
| "step": 92 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 0.000565125, |
| "loss": 1.0857, |
| "step": 93 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 0.0005647499999999999, |
| "loss": 0.8162, |
| "step": 94 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 0.000564375, |
| "loss": 0.9017, |
| "step": 95 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 0.0005639999999999999, |
| "loss": 0.8848, |
| "step": 96 |
| }, |
| { |
| "epoch": 3.03, |
| "learning_rate": 0.0005636249999999999, |
| "loss": 0.9075, |
| "step": 97 |
| }, |
| { |
| "epoch": 3.06, |
| "learning_rate": 0.00056325, |
| "loss": 0.8108, |
| "step": 98 |
| }, |
| { |
| "epoch": 3.09, |
| "learning_rate": 0.0005628749999999999, |
| "loss": 0.8238, |
| "step": 99 |
| }, |
| { |
| "epoch": 3.12, |
| "learning_rate": 0.0005625, |
| "loss": 0.7854, |
| "step": 100 |
| }, |
| { |
| "epoch": 3.12, |
| "eval_loss": 1.0004723072052002, |
| "eval_mean_accuracy": 0.4409761597550485, |
| "eval_mean_iou": 0.3278147775750712, |
| "eval_overall_accuracy": 0.6476280145477831, |
| "eval_per_category_accuracy": [ |
| 0.7394164167070112, |
| 0.7120692919670155, |
| 0.65725755249271, |
| 0.17395090943157923, |
| 0.46355298851110294, |
| 0.31981316619425926, |
| 0.6107480292025965, |
| 0.35138126519456003, |
| 0.1554151176100015, |
| 0.5705962518273815, |
| 0.05702864593695384, |
| 0.4804842819854117 |
| ], |
| "eval_per_category_iou": [ |
| 0.5451244129938286, |
| 0.5178187584431643, |
| 0.5318888189962618, |
| 0.14778251064896017, |
| 0.3169269253463421, |
| 0.2626134690690619, |
| 0.39125714035001896, |
| 0.3081966042250305, |
| 0.12383697911511432, |
| 0.3977603328560864, |
| 0.04942253722870921, |
| 0.34114884162827647 |
| ], |
| "eval_runtime": 124.3127, |
| "eval_samples_per_second": 0.917, |
| "eval_steps_per_second": 0.08, |
| "step": 100 |
| }, |
| { |
| "epoch": 3.16, |
| "learning_rate": 0.0005621249999999999, |
| "loss": 0.8351, |
| "step": 101 |
| }, |
| { |
| "epoch": 3.19, |
| "learning_rate": 0.00056175, |
| "loss": 0.8611, |
| "step": 102 |
| }, |
| { |
| "epoch": 3.22, |
| "learning_rate": 0.000561375, |
| "loss": 0.9022, |
| "step": 103 |
| }, |
| { |
| "epoch": 3.25, |
| "learning_rate": 0.000561, |
| "loss": 0.8189, |
| "step": 104 |
| }, |
| { |
| "epoch": 3.28, |
| "learning_rate": 0.0005606249999999999, |
| "loss": 0.8431, |
| "step": 105 |
| }, |
| { |
| "epoch": 3.31, |
| "learning_rate": 0.00056025, |
| "loss": 0.9712, |
| "step": 106 |
| }, |
| { |
| "epoch": 3.34, |
| "learning_rate": 0.0005598749999999999, |
| "loss": 0.7789, |
| "step": 107 |
| }, |
| { |
| "epoch": 3.38, |
| "learning_rate": 0.0005595, |
| "loss": 0.7784, |
| "step": 108 |
| }, |
| { |
| "epoch": 3.41, |
| "learning_rate": 0.0005591249999999999, |
| "loss": 0.7459, |
| "step": 109 |
| }, |
| { |
| "epoch": 3.44, |
| "learning_rate": 0.00055875, |
| "loss": 0.7602, |
| "step": 110 |
| }, |
| { |
| "epoch": 3.47, |
| "learning_rate": 0.000558375, |
| "loss": 0.7709, |
| "step": 111 |
| }, |
| { |
| "epoch": 3.5, |
| "learning_rate": 0.000558, |
| "loss": 0.792, |
| "step": 112 |
| }, |
| { |
| "epoch": 3.53, |
| "learning_rate": 0.000557625, |
| "loss": 0.7701, |
| "step": 113 |
| }, |
| { |
| "epoch": 3.56, |
| "learning_rate": 0.0005572499999999999, |
| "loss": 0.8632, |
| "step": 114 |
| }, |
| { |
| "epoch": 3.59, |
| "learning_rate": 0.000556875, |
| "loss": 0.776, |
| "step": 115 |
| }, |
| { |
| "epoch": 3.62, |
| "learning_rate": 0.0005564999999999999, |
| "loss": 0.8366, |
| "step": 116 |
| }, |
| { |
| "epoch": 3.66, |
| "learning_rate": 0.000556125, |
| "loss": 0.731, |
| "step": 117 |
| }, |
| { |
| "epoch": 3.69, |
| "learning_rate": 0.0005557499999999999, |
| "loss": 0.8111, |
| "step": 118 |
| }, |
| { |
| "epoch": 3.72, |
| "learning_rate": 0.000555375, |
| "loss": 0.7547, |
| "step": 119 |
| }, |
| { |
| "epoch": 3.75, |
| "learning_rate": 0.0005549999999999999, |
| "loss": 0.7275, |
| "step": 120 |
| }, |
| { |
| "epoch": 3.75, |
| "eval_loss": 1.0055418014526367, |
| "eval_mean_accuracy": 0.46582571564323233, |
| "eval_mean_iou": 0.34071242663610873, |
| "eval_overall_accuracy": 0.6500220717045299, |
| "eval_per_category_accuracy": [ |
| 0.6790541517825218, |
| 0.7248947508922547, |
| 0.702760127940592, |
| 0.22049812284439385, |
| 0.49308863123902213, |
| 0.336202860575091, |
| 0.6458925386011786, |
| 0.40734546148068546, |
| 0.2508473598750699, |
| 0.5798344703996796, |
| 0.12197249573784942, |
| 0.4275176163504493 |
| ], |
| "eval_per_category_iou": [ |
| 0.5331521735997542, |
| 0.5305308321791462, |
| 0.5588320252288795, |
| 0.16697584226263984, |
| 0.3296551780182928, |
| 0.28048869107094193, |
| 0.35449032046775947, |
| 0.34190939733724723, |
| 0.16527723782325413, |
| 0.39484470653333187, |
| 0.0880505368099854, |
| 0.3443421783020718 |
| ], |
| "eval_runtime": 124.4108, |
| "eval_samples_per_second": 0.916, |
| "eval_steps_per_second": 0.08, |
| "step": 120 |
| }, |
| { |
| "epoch": 3.78, |
| "learning_rate": 0.0005546249999999999, |
| "loss": 0.7266, |
| "step": 121 |
| }, |
| { |
| "epoch": 3.81, |
| "learning_rate": 0.00055425, |
| "loss": 0.8377, |
| "step": 122 |
| }, |
| { |
| "epoch": 3.84, |
| "learning_rate": 0.0005538749999999999, |
| "loss": 0.7775, |
| "step": 123 |
| }, |
| { |
| "epoch": 3.88, |
| "learning_rate": 0.0005535, |
| "loss": 0.8343, |
| "step": 124 |
| }, |
| { |
| "epoch": 3.91, |
| "learning_rate": 0.0005531249999999999, |
| "loss": 0.8143, |
| "step": 125 |
| }, |
| { |
| "epoch": 3.94, |
| "learning_rate": 0.00055275, |
| "loss": 0.8377, |
| "step": 126 |
| }, |
| { |
| "epoch": 3.97, |
| "learning_rate": 0.0005523749999999999, |
| "loss": 0.8807, |
| "step": 127 |
| }, |
| { |
| "epoch": 4.0, |
| "learning_rate": 0.000552, |
| "loss": 0.8941, |
| "step": 128 |
| }, |
| { |
| "epoch": 4.03, |
| "learning_rate": 0.000551625, |
| "loss": 0.9077, |
| "step": 129 |
| }, |
| { |
| "epoch": 4.06, |
| "learning_rate": 0.0005512499999999999, |
| "loss": 0.8361, |
| "step": 130 |
| }, |
| { |
| "epoch": 4.09, |
| "learning_rate": 0.0005508749999999999, |
| "loss": 0.6058, |
| "step": 131 |
| }, |
| { |
| "epoch": 4.12, |
| "learning_rate": 0.0005505, |
| "loss": 0.8466, |
| "step": 132 |
| }, |
| { |
| "epoch": 4.16, |
| "learning_rate": 0.0005501249999999999, |
| "loss": 0.6832, |
| "step": 133 |
| }, |
| { |
| "epoch": 4.19, |
| "learning_rate": 0.00054975, |
| "loss": 0.9442, |
| "step": 134 |
| }, |
| { |
| "epoch": 4.22, |
| "learning_rate": 0.000549375, |
| "loss": 0.7119, |
| "step": 135 |
| }, |
| { |
| "epoch": 4.25, |
| "learning_rate": 0.000549, |
| "loss": 0.8412, |
| "step": 136 |
| }, |
| { |
| "epoch": 4.28, |
| "learning_rate": 0.000548625, |
| "loss": 0.8589, |
| "step": 137 |
| }, |
| { |
| "epoch": 4.31, |
| "learning_rate": 0.0005482499999999999, |
| "loss": 0.7529, |
| "step": 138 |
| }, |
| { |
| "epoch": 4.34, |
| "learning_rate": 0.000547875, |
| "loss": 0.7525, |
| "step": 139 |
| }, |
| { |
| "epoch": 4.38, |
| "learning_rate": 0.0005474999999999999, |
| "loss": 0.762, |
| "step": 140 |
| }, |
| { |
| "epoch": 4.38, |
| "eval_loss": 1.009921669960022, |
| "eval_mean_accuracy": 0.4630782466624949, |
| "eval_mean_iou": 0.3357803602667826, |
| "eval_overall_accuracy": 0.6476533454761171, |
| "eval_per_category_accuracy": [ |
| 0.6388404926271997, |
| 0.7725403351834625, |
| 0.6020424381762673, |
| 0.21688954275126474, |
| 0.5520221552773765, |
| 0.4518525508185774, |
| 0.6697900807527053, |
| 0.46810120099302377, |
| 0.1897420654286044, |
| 0.45603045018648025, |
| 0.024152696709344286, |
| 0.5149349510456322 |
| ], |
| "eval_per_category_iou": [ |
| 0.5210483275809908, |
| 0.5385459417489167, |
| 0.519285461653176, |
| 0.19235183880209375, |
| 0.3408901260438495, |
| 0.3345069902216156, |
| 0.36090171027181106, |
| 0.33539228285452394, |
| 0.1422865859237117, |
| 0.3681340241516971, |
| 0.023384620855608634, |
| 0.35263641309339605 |
| ], |
| "eval_runtime": 123.6747, |
| "eval_samples_per_second": 0.922, |
| "eval_steps_per_second": 0.081, |
| "step": 140 |
| }, |
| { |
| "epoch": 4.41, |
| "learning_rate": 0.000547125, |
| "loss": 0.7219, |
| "step": 141 |
| }, |
| { |
| "epoch": 4.44, |
| "learning_rate": 0.0005467499999999999, |
| "loss": 0.7571, |
| "step": 142 |
| }, |
| { |
| "epoch": 4.47, |
| "learning_rate": 0.000546375, |
| "loss": 0.7736, |
| "step": 143 |
| }, |
| { |
| "epoch": 4.5, |
| "learning_rate": 0.0005459999999999999, |
| "loss": 0.7782, |
| "step": 144 |
| }, |
| { |
| "epoch": 4.53, |
| "learning_rate": 0.000545625, |
| "loss": 0.7814, |
| "step": 145 |
| }, |
| { |
| "epoch": 4.56, |
| "learning_rate": 0.0005452499999999999, |
| "loss": 0.6899, |
| "step": 146 |
| }, |
| { |
| "epoch": 4.59, |
| "learning_rate": 0.0005448749999999999, |
| "loss": 0.842, |
| "step": 147 |
| }, |
| { |
| "epoch": 4.62, |
| "learning_rate": 0.0005445, |
| "loss": 0.7052, |
| "step": 148 |
| }, |
| { |
| "epoch": 4.66, |
| "learning_rate": 0.0005441249999999999, |
| "loss": 0.7245, |
| "step": 149 |
| }, |
| { |
| "epoch": 4.69, |
| "learning_rate": 0.00054375, |
| "loss": 0.8852, |
| "step": 150 |
| }, |
| { |
| "epoch": 4.72, |
| "learning_rate": 0.0005433749999999999, |
| "loss": 0.6868, |
| "step": 151 |
| }, |
| { |
| "epoch": 4.75, |
| "learning_rate": 0.000543, |
| "loss": 0.8091, |
| "step": 152 |
| }, |
| { |
| "epoch": 4.78, |
| "learning_rate": 0.000542625, |
| "loss": 0.7912, |
| "step": 153 |
| }, |
| { |
| "epoch": 4.81, |
| "learning_rate": 0.00054225, |
| "loss": 0.8428, |
| "step": 154 |
| }, |
| { |
| "epoch": 4.84, |
| "learning_rate": 0.0005418749999999999, |
| "loss": 0.7627, |
| "step": 155 |
| }, |
| { |
| "epoch": 4.88, |
| "learning_rate": 0.0005414999999999999, |
| "loss": 0.6705, |
| "step": 156 |
| }, |
| { |
| "epoch": 4.91, |
| "learning_rate": 0.0005411249999999999, |
| "loss": 0.7269, |
| "step": 157 |
| }, |
| { |
| "epoch": 4.94, |
| "learning_rate": 0.00054075, |
| "loss": 0.7688, |
| "step": 158 |
| }, |
| { |
| "epoch": 4.97, |
| "learning_rate": 0.000540375, |
| "loss": 0.757, |
| "step": 159 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 0.00054, |
| "loss": 0.7543, |
| "step": 160 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_loss": 0.9964900612831116, |
| "eval_mean_accuracy": 0.47220927209560704, |
| "eval_mean_iou": 0.35076273333971525, |
| "eval_overall_accuracy": 0.6553197827255517, |
| "eval_per_category_accuracy": [ |
| 0.5974941039770494, |
| 0.7656300402304882, |
| 0.7736861734503012, |
| 0.26844019125474494, |
| 0.47782547654307983, |
| 0.5049230780397693, |
| 0.5503440924885342, |
| 0.38577494532904893, |
| 0.29696414469889887, |
| 0.6231179209340338, |
| 0.08607248563724125, |
| 0.33623861256409465 |
| ], |
| "eval_per_category_iou": [ |
| 0.5029245432853789, |
| 0.5438264868522664, |
| 0.5504928848830769, |
| 0.20921503460100002, |
| 0.3484194245550177, |
| 0.34286066341367244, |
| 0.4044913460614946, |
| 0.3442055933600121, |
| 0.19153018129377833, |
| 0.4114387452368694, |
| 0.07501013614442405, |
| 0.28473776038959164 |
| ], |
| "eval_runtime": 123.8413, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 160 |
| }, |
| { |
| "epoch": 5.03, |
| "learning_rate": 0.000539625, |
| "loss": 0.639, |
| "step": 161 |
| }, |
| { |
| "epoch": 5.06, |
| "learning_rate": 0.00053925, |
| "loss": 0.7108, |
| "step": 162 |
| }, |
| { |
| "epoch": 5.09, |
| "learning_rate": 0.000538875, |
| "loss": 0.6735, |
| "step": 163 |
| }, |
| { |
| "epoch": 5.12, |
| "learning_rate": 0.0005384999999999999, |
| "loss": 0.7478, |
| "step": 164 |
| }, |
| { |
| "epoch": 5.16, |
| "learning_rate": 0.000538125, |
| "loss": 0.6023, |
| "step": 165 |
| }, |
| { |
| "epoch": 5.19, |
| "learning_rate": 0.0005377499999999999, |
| "loss": 0.7976, |
| "step": 166 |
| }, |
| { |
| "epoch": 5.22, |
| "learning_rate": 0.000537375, |
| "loss": 0.7348, |
| "step": 167 |
| }, |
| { |
| "epoch": 5.25, |
| "learning_rate": 0.0005369999999999999, |
| "loss": 0.7037, |
| "step": 168 |
| }, |
| { |
| "epoch": 5.28, |
| "learning_rate": 0.000536625, |
| "loss": 0.694, |
| "step": 169 |
| }, |
| { |
| "epoch": 5.31, |
| "learning_rate": 0.0005362499999999999, |
| "loss": 0.7427, |
| "step": 170 |
| }, |
| { |
| "epoch": 5.34, |
| "learning_rate": 0.0005358749999999999, |
| "loss": 0.7319, |
| "step": 171 |
| }, |
| { |
| "epoch": 5.38, |
| "learning_rate": 0.0005355, |
| "loss": 0.7365, |
| "step": 172 |
| }, |
| { |
| "epoch": 5.41, |
| "learning_rate": 0.0005351249999999999, |
| "loss": 0.823, |
| "step": 173 |
| }, |
| { |
| "epoch": 5.44, |
| "learning_rate": 0.00053475, |
| "loss": 0.6932, |
| "step": 174 |
| }, |
| { |
| "epoch": 5.47, |
| "learning_rate": 0.0005343749999999999, |
| "loss": 0.6911, |
| "step": 175 |
| }, |
| { |
| "epoch": 5.5, |
| "learning_rate": 0.000534, |
| "loss": 0.7337, |
| "step": 176 |
| }, |
| { |
| "epoch": 5.53, |
| "learning_rate": 0.0005336249999999999, |
| "loss": 0.8752, |
| "step": 177 |
| }, |
| { |
| "epoch": 5.56, |
| "learning_rate": 0.00053325, |
| "loss": 0.669, |
| "step": 178 |
| }, |
| { |
| "epoch": 5.59, |
| "learning_rate": 0.000532875, |
| "loss": 0.7305, |
| "step": 179 |
| }, |
| { |
| "epoch": 5.62, |
| "learning_rate": 0.0005324999999999999, |
| "loss": 0.7142, |
| "step": 180 |
| }, |
| { |
| "epoch": 5.62, |
| "eval_loss": 0.9814519286155701, |
| "eval_mean_accuracy": 0.497081755026592, |
| "eval_mean_iou": 0.35939949129333554, |
| "eval_overall_accuracy": 0.6597169909560889, |
| "eval_per_category_accuracy": [ |
| 0.6908915714734837, |
| 0.7305051261897875, |
| 0.7036370157962516, |
| 0.2776663590120818, |
| 0.5628060380018922, |
| 0.4295562515651142, |
| 0.5859168478577468, |
| 0.2814967358147499, |
| 0.3033792356902656, |
| 0.6404991026728926, |
| 0.30056349150477635, |
| 0.4580632847400621 |
| ], |
| "eval_per_category_iou": [ |
| 0.5553982956913027, |
| 0.5378390398854487, |
| 0.5530532106585494, |
| 0.21904999137749334, |
| 0.3437066198831567, |
| 0.32814516520304693, |
| 0.4234087095030862, |
| 0.2666791179990438, |
| 0.18338491161074402, |
| 0.41252681863698343, |
| 0.14896654252535288, |
| 0.34063547254581833 |
| ], |
| "eval_runtime": 124.8537, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 180 |
| }, |
| { |
| "epoch": 5.66, |
| "learning_rate": 0.0005321249999999999, |
| "loss": 0.7298, |
| "step": 181 |
| }, |
| { |
| "epoch": 5.69, |
| "learning_rate": 0.00053175, |
| "loss": 0.6999, |
| "step": 182 |
| }, |
| { |
| "epoch": 5.72, |
| "learning_rate": 0.000531375, |
| "loss": 0.8023, |
| "step": 183 |
| }, |
| { |
| "epoch": 5.75, |
| "learning_rate": 0.000531, |
| "loss": 0.6922, |
| "step": 184 |
| }, |
| { |
| "epoch": 5.78, |
| "learning_rate": 0.000530625, |
| "loss": 0.7184, |
| "step": 185 |
| }, |
| { |
| "epoch": 5.81, |
| "learning_rate": 0.00053025, |
| "loss": 0.7189, |
| "step": 186 |
| }, |
| { |
| "epoch": 5.84, |
| "learning_rate": 0.000529875, |
| "loss": 0.614, |
| "step": 187 |
| }, |
| { |
| "epoch": 5.88, |
| "learning_rate": 0.0005294999999999999, |
| "loss": 0.6966, |
| "step": 188 |
| }, |
| { |
| "epoch": 5.91, |
| "learning_rate": 0.000529125, |
| "loss": 0.7259, |
| "step": 189 |
| }, |
| { |
| "epoch": 5.94, |
| "learning_rate": 0.0005287499999999999, |
| "loss": 0.7856, |
| "step": 190 |
| }, |
| { |
| "epoch": 5.97, |
| "learning_rate": 0.000528375, |
| "loss": 0.7678, |
| "step": 191 |
| }, |
| { |
| "epoch": 6.0, |
| "learning_rate": 0.0005279999999999999, |
| "loss": 0.623, |
| "step": 192 |
| }, |
| { |
| "epoch": 6.03, |
| "learning_rate": 0.000527625, |
| "loss": 0.6785, |
| "step": 193 |
| }, |
| { |
| "epoch": 6.06, |
| "learning_rate": 0.0005272499999999999, |
| "loss": 0.6884, |
| "step": 194 |
| }, |
| { |
| "epoch": 6.09, |
| "learning_rate": 0.000526875, |
| "loss": 0.6096, |
| "step": 195 |
| }, |
| { |
| "epoch": 6.12, |
| "learning_rate": 0.0005265, |
| "loss": 0.739, |
| "step": 196 |
| }, |
| { |
| "epoch": 6.16, |
| "learning_rate": 0.0005261249999999999, |
| "loss": 0.6649, |
| "step": 197 |
| }, |
| { |
| "epoch": 6.19, |
| "learning_rate": 0.00052575, |
| "loss": 0.6959, |
| "step": 198 |
| }, |
| { |
| "epoch": 6.22, |
| "learning_rate": 0.0005253749999999999, |
| "loss": 0.6718, |
| "step": 199 |
| }, |
| { |
| "epoch": 6.25, |
| "learning_rate": 0.000525, |
| "loss": 0.7132, |
| "step": 200 |
| }, |
| { |
| "epoch": 6.25, |
| "eval_loss": 0.9577063918113708, |
| "eval_mean_accuracy": 0.49358942386599547, |
| "eval_mean_iou": 0.36445855775079217, |
| "eval_overall_accuracy": 0.6702703174791838, |
| "eval_per_category_accuracy": [ |
| 0.757485170186219, |
| 0.7039742826269, |
| 0.7879783326232815, |
| 0.3002650918453029, |
| 0.530024898350307, |
| 0.5117678441164408, |
| 0.5200409667652787, |
| 0.282272468662643, |
| 0.18195167830492534, |
| 0.6273017965839506, |
| 0.20833269566867557, |
| 0.511677860658021 |
| ], |
| "eval_per_category_iou": [ |
| 0.5717339461326115, |
| 0.5411086678260705, |
| 0.5550507730528956, |
| 0.23120184241180683, |
| 0.33923086199004965, |
| 0.35084943122443496, |
| 0.4183980452164044, |
| 0.2680651542162202, |
| 0.15420450059218319, |
| 0.4336026908109024, |
| 0.15032155761093297, |
| 0.3597352219249941 |
| ], |
| "eval_runtime": 123.8598, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 200 |
| }, |
| { |
| "epoch": 6.28, |
| "learning_rate": 0.0005246249999999999, |
| "loss": 0.661, |
| "step": 201 |
| }, |
| { |
| "epoch": 6.31, |
| "learning_rate": 0.00052425, |
| "loss": 0.7055, |
| "step": 202 |
| }, |
| { |
| "epoch": 6.34, |
| "learning_rate": 0.0005238749999999999, |
| "loss": 0.6852, |
| "step": 203 |
| }, |
| { |
| "epoch": 6.38, |
| "learning_rate": 0.0005235, |
| "loss": 0.6802, |
| "step": 204 |
| }, |
| { |
| "epoch": 6.41, |
| "learning_rate": 0.0005231249999999999, |
| "loss": 0.6662, |
| "step": 205 |
| }, |
| { |
| "epoch": 6.44, |
| "learning_rate": 0.0005227499999999999, |
| "loss": 0.7165, |
| "step": 206 |
| }, |
| { |
| "epoch": 6.47, |
| "learning_rate": 0.0005223749999999999, |
| "loss": 0.6282, |
| "step": 207 |
| }, |
| { |
| "epoch": 6.5, |
| "learning_rate": 0.000522, |
| "loss": 0.693, |
| "step": 208 |
| }, |
| { |
| "epoch": 6.53, |
| "learning_rate": 0.000521625, |
| "loss": 0.6498, |
| "step": 209 |
| }, |
| { |
| "epoch": 6.56, |
| "learning_rate": 0.00052125, |
| "loss": 0.7024, |
| "step": 210 |
| }, |
| { |
| "epoch": 6.59, |
| "learning_rate": 0.000520875, |
| "loss": 0.573, |
| "step": 211 |
| }, |
| { |
| "epoch": 6.62, |
| "learning_rate": 0.0005205, |
| "loss": 0.8226, |
| "step": 212 |
| }, |
| { |
| "epoch": 6.66, |
| "learning_rate": 0.000520125, |
| "loss": 0.5298, |
| "step": 213 |
| }, |
| { |
| "epoch": 6.69, |
| "learning_rate": 0.0005197499999999999, |
| "loss": 0.6492, |
| "step": 214 |
| }, |
| { |
| "epoch": 6.72, |
| "learning_rate": 0.000519375, |
| "loss": 0.769, |
| "step": 215 |
| }, |
| { |
| "epoch": 6.75, |
| "learning_rate": 0.0005189999999999999, |
| "loss": 0.664, |
| "step": 216 |
| }, |
| { |
| "epoch": 6.78, |
| "learning_rate": 0.000518625, |
| "loss": 0.7115, |
| "step": 217 |
| }, |
| { |
| "epoch": 6.81, |
| "learning_rate": 0.0005182499999999999, |
| "loss": 0.6526, |
| "step": 218 |
| }, |
| { |
| "epoch": 6.84, |
| "learning_rate": 0.000517875, |
| "loss": 0.679, |
| "step": 219 |
| }, |
| { |
| "epoch": 6.88, |
| "learning_rate": 0.0005175, |
| "loss": 0.5572, |
| "step": 220 |
| }, |
| { |
| "epoch": 6.88, |
| "eval_loss": 0.9686026573181152, |
| "eval_mean_accuracy": 0.49162090929743196, |
| "eval_mean_iou": 0.3674592922271513, |
| "eval_overall_accuracy": 0.668428052935684, |
| "eval_per_category_accuracy": [ |
| 0.8315143436240795, |
| 0.67630998091554, |
| 0.7426204308307264, |
| 0.32498386548323743, |
| 0.49111945333028095, |
| 0.5001669455140652, |
| 0.5374337360209966, |
| 0.43089816792044666, |
| 0.34422109438080617, |
| 0.6128100029154656, |
| 0.11920860204522012, |
| 0.28816428858831905 |
| ], |
| "eval_per_category_iou": [ |
| 0.5778473653822723, |
| 0.5269478191058978, |
| 0.574435506709114, |
| 0.23668203929496492, |
| 0.3570071601716556, |
| 0.358769802685785, |
| 0.41459541328306615, |
| 0.37143503659747396, |
| 0.2224686465997259, |
| 0.4152571514805167, |
| 0.09885903570114096, |
| 0.25520652971420166 |
| ], |
| "eval_runtime": 124.6105, |
| "eval_samples_per_second": 0.915, |
| "eval_steps_per_second": 0.08, |
| "step": 220 |
| }, |
| { |
| "epoch": 6.91, |
| "learning_rate": 0.0005171249999999999, |
| "loss": 0.7046, |
| "step": 221 |
| }, |
| { |
| "epoch": 6.94, |
| "learning_rate": 0.00051675, |
| "loss": 0.6436, |
| "step": 222 |
| }, |
| { |
| "epoch": 6.97, |
| "learning_rate": 0.0005163749999999999, |
| "loss": 0.6286, |
| "step": 223 |
| }, |
| { |
| "epoch": 7.0, |
| "learning_rate": 0.000516, |
| "loss": 0.621, |
| "step": 224 |
| }, |
| { |
| "epoch": 7.03, |
| "learning_rate": 0.0005156249999999999, |
| "loss": 0.7192, |
| "step": 225 |
| }, |
| { |
| "epoch": 7.06, |
| "learning_rate": 0.00051525, |
| "loss": 0.5768, |
| "step": 226 |
| }, |
| { |
| "epoch": 7.09, |
| "learning_rate": 0.0005148749999999999, |
| "loss": 0.6557, |
| "step": 227 |
| }, |
| { |
| "epoch": 7.12, |
| "learning_rate": 0.0005145, |
| "loss": 0.6194, |
| "step": 228 |
| }, |
| { |
| "epoch": 7.16, |
| "learning_rate": 0.0005141249999999999, |
| "loss": 0.624, |
| "step": 229 |
| }, |
| { |
| "epoch": 7.19, |
| "learning_rate": 0.0005137499999999999, |
| "loss": 0.6787, |
| "step": 230 |
| }, |
| { |
| "epoch": 7.22, |
| "learning_rate": 0.0005133749999999999, |
| "loss": 0.6515, |
| "step": 231 |
| }, |
| { |
| "epoch": 7.25, |
| "learning_rate": 0.0005129999999999999, |
| "loss": 0.618, |
| "step": 232 |
| }, |
| { |
| "epoch": 7.28, |
| "learning_rate": 0.000512625, |
| "loss": 0.5743, |
| "step": 233 |
| }, |
| { |
| "epoch": 7.31, |
| "learning_rate": 0.00051225, |
| "loss": 0.561, |
| "step": 234 |
| }, |
| { |
| "epoch": 7.34, |
| "learning_rate": 0.000511875, |
| "loss": 0.656, |
| "step": 235 |
| }, |
| { |
| "epoch": 7.38, |
| "learning_rate": 0.0005115, |
| "loss": 0.5951, |
| "step": 236 |
| }, |
| { |
| "epoch": 7.41, |
| "learning_rate": 0.000511125, |
| "loss": 0.6536, |
| "step": 237 |
| }, |
| { |
| "epoch": 7.44, |
| "learning_rate": 0.0005107499999999999, |
| "loss": 0.734, |
| "step": 238 |
| }, |
| { |
| "epoch": 7.47, |
| "learning_rate": 0.000510375, |
| "loss": 0.6387, |
| "step": 239 |
| }, |
| { |
| "epoch": 7.5, |
| "learning_rate": 0.0005099999999999999, |
| "loss": 0.5676, |
| "step": 240 |
| }, |
| { |
| "epoch": 7.5, |
| "eval_loss": 0.9805742502212524, |
| "eval_mean_accuracy": 0.5024018526714576, |
| "eval_mean_iou": 0.371506102641304, |
| "eval_overall_accuracy": 0.6800810161389803, |
| "eval_per_category_accuracy": [ |
| 0.7315930355784842, |
| 0.7541007820537842, |
| 0.6881292554956727, |
| 0.30834622938078154, |
| 0.5766004128790718, |
| 0.4276563390300466, |
| 0.6505111399057544, |
| 0.4819251032078268, |
| 0.27926275254429783, |
| 0.5722067988240086, |
| 0.07342223909057798, |
| 0.48506814406718457 |
| ], |
| "eval_per_category_iou": [ |
| 0.5874312271992882, |
| 0.5604905946583346, |
| 0.5587517085148045, |
| 0.2254583833345596, |
| 0.36177477005679304, |
| 0.33678157470165493, |
| 0.4185254872962371, |
| 0.3728270419920094, |
| 0.20258071367884453, |
| 0.42075932706243124, |
| 0.0692206473059698, |
| 0.34347175589472057 |
| ], |
| "eval_runtime": 123.7241, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 240 |
| }, |
| { |
| "epoch": 7.53, |
| "learning_rate": 0.000509625, |
| "loss": 0.6097, |
| "step": 241 |
| }, |
| { |
| "epoch": 7.56, |
| "learning_rate": 0.0005092499999999999, |
| "loss": 0.6319, |
| "step": 242 |
| }, |
| { |
| "epoch": 7.59, |
| "learning_rate": 0.000508875, |
| "loss": 0.6219, |
| "step": 243 |
| }, |
| { |
| "epoch": 7.62, |
| "learning_rate": 0.0005085, |
| "loss": 0.5975, |
| "step": 244 |
| }, |
| { |
| "epoch": 7.66, |
| "learning_rate": 0.000508125, |
| "loss": 0.651, |
| "step": 245 |
| }, |
| { |
| "epoch": 7.69, |
| "learning_rate": 0.00050775, |
| "loss": 0.512, |
| "step": 246 |
| }, |
| { |
| "epoch": 7.72, |
| "learning_rate": 0.0005073749999999999, |
| "loss": 0.5962, |
| "step": 247 |
| }, |
| { |
| "epoch": 7.75, |
| "learning_rate": 0.000507, |
| "loss": 0.7017, |
| "step": 248 |
| }, |
| { |
| "epoch": 7.78, |
| "learning_rate": 0.0005066249999999999, |
| "loss": 0.6477, |
| "step": 249 |
| }, |
| { |
| "epoch": 7.81, |
| "learning_rate": 0.00050625, |
| "loss": 0.6546, |
| "step": 250 |
| }, |
| { |
| "epoch": 7.84, |
| "learning_rate": 0.0005058749999999999, |
| "loss": 0.6881, |
| "step": 251 |
| }, |
| { |
| "epoch": 7.88, |
| "learning_rate": 0.0005055, |
| "loss": 0.6078, |
| "step": 252 |
| }, |
| { |
| "epoch": 7.91, |
| "learning_rate": 0.0005051249999999999, |
| "loss": 0.6891, |
| "step": 253 |
| }, |
| { |
| "epoch": 7.94, |
| "learning_rate": 0.00050475, |
| "loss": 0.598, |
| "step": 254 |
| }, |
| { |
| "epoch": 7.97, |
| "learning_rate": 0.0005043749999999999, |
| "loss": 0.638, |
| "step": 255 |
| }, |
| { |
| "epoch": 8.0, |
| "learning_rate": 0.0005039999999999999, |
| "loss": 0.6779, |
| "step": 256 |
| }, |
| { |
| "epoch": 8.03, |
| "learning_rate": 0.000503625, |
| "loss": 0.628, |
| "step": 257 |
| }, |
| { |
| "epoch": 8.06, |
| "learning_rate": 0.0005032499999999999, |
| "loss": 0.5842, |
| "step": 258 |
| }, |
| { |
| "epoch": 8.09, |
| "learning_rate": 0.000502875, |
| "loss": 0.5126, |
| "step": 259 |
| }, |
| { |
| "epoch": 8.12, |
| "learning_rate": 0.0005025, |
| "loss": 0.6533, |
| "step": 260 |
| }, |
| { |
| "epoch": 8.12, |
| "eval_loss": 0.9647707939147949, |
| "eval_mean_accuracy": 0.5290347832309982, |
| "eval_mean_iou": 0.3826944916955128, |
| "eval_overall_accuracy": 0.6695300654361123, |
| "eval_per_category_accuracy": [ |
| 0.6390227219897978, |
| 0.7622451073333125, |
| 0.6798905616562309, |
| 0.39068084190949404, |
| 0.5202418118057482, |
| 0.5864632592845296, |
| 0.7008623173460086, |
| 0.5284176442793452, |
| 0.26142604899603383, |
| 0.5910629245092603, |
| 0.23104988170045268, |
| 0.45705427796176507 |
| ], |
| "eval_per_category_iou": [ |
| 0.539363765599418, |
| 0.5541112560719536, |
| 0.564224130990146, |
| 0.28007313069001544, |
| 0.35043846601035505, |
| 0.37497345958257, |
| 0.4145645950481569, |
| 0.40213470319634703, |
| 0.1947903409748277, |
| 0.4211198117656152, |
| 0.15867080332782618, |
| 0.33786943708892114 |
| ], |
| "eval_runtime": 125.2253, |
| "eval_samples_per_second": 0.91, |
| "eval_steps_per_second": 0.08, |
| "step": 260 |
| }, |
| { |
| "epoch": 8.16, |
| "learning_rate": 0.000502125, |
| "loss": 0.5969, |
| "step": 261 |
| }, |
| { |
| "epoch": 8.19, |
| "learning_rate": 0.00050175, |
| "loss": 0.5633, |
| "step": 262 |
| }, |
| { |
| "epoch": 8.22, |
| "learning_rate": 0.000501375, |
| "loss": 0.6202, |
| "step": 263 |
| }, |
| { |
| "epoch": 8.25, |
| "learning_rate": 0.0005009999999999999, |
| "loss": 0.5568, |
| "step": 264 |
| }, |
| { |
| "epoch": 8.28, |
| "learning_rate": 0.000500625, |
| "loss": 0.6021, |
| "step": 265 |
| }, |
| { |
| "epoch": 8.31, |
| "learning_rate": 0.0005002499999999999, |
| "loss": 0.6637, |
| "step": 266 |
| }, |
| { |
| "epoch": 8.34, |
| "learning_rate": 0.000499875, |
| "loss": 0.6756, |
| "step": 267 |
| }, |
| { |
| "epoch": 8.38, |
| "learning_rate": 0.0004994999999999999, |
| "loss": 0.6323, |
| "step": 268 |
| }, |
| { |
| "epoch": 8.41, |
| "learning_rate": 0.000499125, |
| "loss": 0.6591, |
| "step": 269 |
| }, |
| { |
| "epoch": 8.44, |
| "learning_rate": 0.00049875, |
| "loss": 0.7191, |
| "step": 270 |
| }, |
| { |
| "epoch": 8.47, |
| "learning_rate": 0.0004983749999999999, |
| "loss": 0.5911, |
| "step": 271 |
| }, |
| { |
| "epoch": 8.5, |
| "learning_rate": 0.000498, |
| "loss": 0.5841, |
| "step": 272 |
| }, |
| { |
| "epoch": 8.53, |
| "learning_rate": 0.0004976249999999999, |
| "loss": 0.5837, |
| "step": 273 |
| }, |
| { |
| "epoch": 8.56, |
| "learning_rate": 0.00049725, |
| "loss": 0.6009, |
| "step": 274 |
| }, |
| { |
| "epoch": 8.59, |
| "learning_rate": 0.0004968749999999999, |
| "loss": 0.5528, |
| "step": 275 |
| }, |
| { |
| "epoch": 8.62, |
| "learning_rate": 0.0004965, |
| "loss": 0.5907, |
| "step": 276 |
| }, |
| { |
| "epoch": 8.66, |
| "learning_rate": 0.0004961249999999999, |
| "loss": 0.5582, |
| "step": 277 |
| }, |
| { |
| "epoch": 8.69, |
| "learning_rate": 0.00049575, |
| "loss": 0.5714, |
| "step": 278 |
| }, |
| { |
| "epoch": 8.72, |
| "learning_rate": 0.0004953749999999999, |
| "loss": 0.6407, |
| "step": 279 |
| }, |
| { |
| "epoch": 8.75, |
| "learning_rate": 0.0004949999999999999, |
| "loss": 0.67, |
| "step": 280 |
| }, |
| { |
| "epoch": 8.75, |
| "eval_loss": 1.0987780094146729, |
| "eval_mean_accuracy": 0.4580998870390869, |
| "eval_mean_iou": 0.3479111276469824, |
| "eval_overall_accuracy": 0.6546705814830044, |
| "eval_per_category_accuracy": [ |
| 0.48066672745873446, |
| 0.8305757369135304, |
| 0.7676997659254449, |
| 0.21412412127604943, |
| 0.47398841588551216, |
| 0.4325884901955803, |
| 0.5711691272441585, |
| 0.2749094353329169, |
| 0.21508300577624376, |
| 0.6848772870627909, |
| 0.20816741298938213, |
| 0.3433491184086993 |
| ], |
| "eval_per_category_iou": [ |
| 0.4455015326675329, |
| 0.5573106019694799, |
| 0.5610375857814821, |
| 0.19477330471696624, |
| 0.35352567686407715, |
| 0.32306020230054533, |
| 0.4240965616003975, |
| 0.26065154759824616, |
| 0.18418258317583033, |
| 0.4215337886509578, |
| 0.15863326887955384, |
| 0.2906268775587196 |
| ], |
| "eval_runtime": 123.9472, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 280 |
| }, |
| { |
| "epoch": 8.78, |
| "learning_rate": 0.000494625, |
| "loss": 0.6267, |
| "step": 281 |
| }, |
| { |
| "epoch": 8.81, |
| "learning_rate": 0.0004942499999999999, |
| "loss": 0.7083, |
| "step": 282 |
| }, |
| { |
| "epoch": 8.84, |
| "learning_rate": 0.000493875, |
| "loss": 0.6574, |
| "step": 283 |
| }, |
| { |
| "epoch": 8.88, |
| "learning_rate": 0.0004935, |
| "loss": 0.5851, |
| "step": 284 |
| }, |
| { |
| "epoch": 8.91, |
| "learning_rate": 0.000493125, |
| "loss": 0.5928, |
| "step": 285 |
| }, |
| { |
| "epoch": 8.94, |
| "learning_rate": 0.00049275, |
| "loss": 0.5435, |
| "step": 286 |
| }, |
| { |
| "epoch": 8.97, |
| "learning_rate": 0.000492375, |
| "loss": 0.6518, |
| "step": 287 |
| }, |
| { |
| "epoch": 9.0, |
| "learning_rate": 0.0004919999999999999, |
| "loss": 0.654, |
| "step": 288 |
| }, |
| { |
| "epoch": 9.03, |
| "learning_rate": 0.000491625, |
| "loss": 0.5646, |
| "step": 289 |
| }, |
| { |
| "epoch": 9.06, |
| "learning_rate": 0.0004912499999999999, |
| "loss": 0.5048, |
| "step": 290 |
| }, |
| { |
| "epoch": 9.09, |
| "learning_rate": 0.000490875, |
| "loss": 0.5574, |
| "step": 291 |
| }, |
| { |
| "epoch": 9.12, |
| "learning_rate": 0.0004904999999999999, |
| "loss": 0.5995, |
| "step": 292 |
| }, |
| { |
| "epoch": 9.16, |
| "learning_rate": 0.000490125, |
| "loss": 0.5645, |
| "step": 293 |
| }, |
| { |
| "epoch": 9.19, |
| "learning_rate": 0.00048975, |
| "loss": 0.5597, |
| "step": 294 |
| }, |
| { |
| "epoch": 9.22, |
| "learning_rate": 0.000489375, |
| "loss": 0.6295, |
| "step": 295 |
| }, |
| { |
| "epoch": 9.25, |
| "learning_rate": 0.000489, |
| "loss": 0.5053, |
| "step": 296 |
| }, |
| { |
| "epoch": 9.28, |
| "learning_rate": 0.0004886249999999999, |
| "loss": 0.4657, |
| "step": 297 |
| }, |
| { |
| "epoch": 9.31, |
| "learning_rate": 0.0004882499999999999, |
| "loss": 0.6421, |
| "step": 298 |
| }, |
| { |
| "epoch": 9.34, |
| "learning_rate": 0.000487875, |
| "loss": 0.5838, |
| "step": 299 |
| }, |
| { |
| "epoch": 9.38, |
| "learning_rate": 0.0004875, |
| "loss": 0.5644, |
| "step": 300 |
| }, |
| { |
| "epoch": 9.38, |
| "eval_loss": 0.983669638633728, |
| "eval_mean_accuracy": 0.5204264301169051, |
| "eval_mean_iou": 0.36998021455594227, |
| "eval_overall_accuracy": 0.6649547710753324, |
| "eval_per_category_accuracy": [ |
| 0.5940242572097704, |
| 0.8056204971919827, |
| 0.652415122648732, |
| 0.3024337096897315, |
| 0.6671805707779671, |
| 0.7223677954844868, |
| 0.5593419811729217, |
| 0.49315930059755003, |
| 0.35564048871813525, |
| 0.4758290170708594, |
| 0.1260647724455409, |
| 0.491039648395184 |
| ], |
| "eval_per_category_iou": [ |
| 0.5120521418371993, |
| 0.5615051709697225, |
| 0.5605287190343511, |
| 0.24633585240536524, |
| 0.31312694322687384, |
| 0.3423289627909097, |
| 0.42465141853084787, |
| 0.4075806091690605, |
| 0.2081047227792813, |
| 0.3872001100058186, |
| 0.10902319327019286, |
| 0.36732473065168403 |
| ], |
| "eval_runtime": 123.8999, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 300 |
| }, |
| { |
| "epoch": 9.41, |
| "learning_rate": 0.000487125, |
| "loss": 0.6492, |
| "step": 301 |
| }, |
| { |
| "epoch": 9.44, |
| "learning_rate": 0.00048675, |
| "loss": 0.6658, |
| "step": 302 |
| }, |
| { |
| "epoch": 9.47, |
| "learning_rate": 0.000486375, |
| "loss": 0.5424, |
| "step": 303 |
| }, |
| { |
| "epoch": 9.5, |
| "learning_rate": 0.000486, |
| "loss": 0.4951, |
| "step": 304 |
| }, |
| { |
| "epoch": 9.53, |
| "learning_rate": 0.00048562499999999995, |
| "loss": 0.5663, |
| "step": 305 |
| }, |
| { |
| "epoch": 9.56, |
| "learning_rate": 0.00048524999999999995, |
| "loss": 0.5105, |
| "step": 306 |
| }, |
| { |
| "epoch": 9.59, |
| "learning_rate": 0.00048487499999999995, |
| "loss": 0.5594, |
| "step": 307 |
| }, |
| { |
| "epoch": 9.62, |
| "learning_rate": 0.00048449999999999996, |
| "loss": 0.5925, |
| "step": 308 |
| }, |
| { |
| "epoch": 9.66, |
| "learning_rate": 0.00048412499999999996, |
| "loss": 0.5676, |
| "step": 309 |
| }, |
| { |
| "epoch": 9.69, |
| "learning_rate": 0.00048374999999999997, |
| "loss": 0.552, |
| "step": 310 |
| }, |
| { |
| "epoch": 9.72, |
| "learning_rate": 0.00048337499999999997, |
| "loss": 0.6232, |
| "step": 311 |
| }, |
| { |
| "epoch": 9.75, |
| "learning_rate": 0.000483, |
| "loss": 0.5877, |
| "step": 312 |
| }, |
| { |
| "epoch": 9.78, |
| "learning_rate": 0.0004826249999999999, |
| "loss": 0.5313, |
| "step": 313 |
| }, |
| { |
| "epoch": 9.81, |
| "learning_rate": 0.00048224999999999993, |
| "loss": 0.5456, |
| "step": 314 |
| }, |
| { |
| "epoch": 9.84, |
| "learning_rate": 0.00048187499999999994, |
| "loss": 0.5813, |
| "step": 315 |
| }, |
| { |
| "epoch": 9.88, |
| "learning_rate": 0.00048149999999999994, |
| "loss": 0.6503, |
| "step": 316 |
| }, |
| { |
| "epoch": 9.91, |
| "learning_rate": 0.00048112499999999994, |
| "loss": 0.5662, |
| "step": 317 |
| }, |
| { |
| "epoch": 9.94, |
| "learning_rate": 0.00048074999999999995, |
| "loss": 0.6718, |
| "step": 318 |
| }, |
| { |
| "epoch": 9.97, |
| "learning_rate": 0.00048037499999999995, |
| "loss": 0.5636, |
| "step": 319 |
| }, |
| { |
| "epoch": 10.0, |
| "learning_rate": 0.00047999999999999996, |
| "loss": 0.5678, |
| "step": 320 |
| }, |
| { |
| "epoch": 10.0, |
| "eval_loss": 0.9780566692352295, |
| "eval_mean_accuracy": 0.5350595191545178, |
| "eval_mean_iou": 0.3906962436607775, |
| "eval_overall_accuracy": 0.6771719413891173, |
| "eval_per_category_accuracy": [ |
| 0.6554152883827387, |
| 0.7588004293882252, |
| 0.7401698246138683, |
| 0.3209831992838356, |
| 0.5589183297643672, |
| 0.5685601675261941, |
| 0.5751910181705395, |
| 0.45440587328339643, |
| 0.4784721790901751, |
| 0.6453562996222412, |
| 0.21756097859589305, |
| 0.4468806421327391 |
| ], |
| "eval_per_category_iou": [ |
| 0.5545401246305556, |
| 0.559287046180608, |
| 0.5841953654092598, |
| 0.24684268808453297, |
| 0.3812876262500622, |
| 0.3824948025455624, |
| 0.4319588439462186, |
| 0.384631151257913, |
| 0.2309813395356408, |
| 0.4207958788984324, |
| 0.15715858689932388, |
| 0.3541814702912209 |
| ], |
| "eval_runtime": 123.6926, |
| "eval_samples_per_second": 0.922, |
| "eval_steps_per_second": 0.081, |
| "step": 320 |
| }, |
| { |
| "epoch": 10.03, |
| "learning_rate": 0.0004796249999999999, |
| "loss": 0.5511, |
| "step": 321 |
| }, |
| { |
| "epoch": 10.06, |
| "learning_rate": 0.0004792499999999999, |
| "loss": 0.611, |
| "step": 322 |
| }, |
| { |
| "epoch": 10.09, |
| "learning_rate": 0.0004788749999999999, |
| "loss": 0.5803, |
| "step": 323 |
| }, |
| { |
| "epoch": 10.12, |
| "learning_rate": 0.0004785, |
| "loss": 0.5232, |
| "step": 324 |
| }, |
| { |
| "epoch": 10.16, |
| "learning_rate": 0.000478125, |
| "loss": 0.5558, |
| "step": 325 |
| }, |
| { |
| "epoch": 10.19, |
| "learning_rate": 0.00047775, |
| "loss": 0.5829, |
| "step": 326 |
| }, |
| { |
| "epoch": 10.22, |
| "learning_rate": 0.000477375, |
| "loss": 0.5082, |
| "step": 327 |
| }, |
| { |
| "epoch": 10.25, |
| "learning_rate": 0.000477, |
| "loss": 0.5775, |
| "step": 328 |
| }, |
| { |
| "epoch": 10.28, |
| "learning_rate": 0.000476625, |
| "loss": 0.5354, |
| "step": 329 |
| }, |
| { |
| "epoch": 10.31, |
| "learning_rate": 0.00047624999999999995, |
| "loss": 0.4546, |
| "step": 330 |
| }, |
| { |
| "epoch": 10.34, |
| "learning_rate": 0.00047587499999999995, |
| "loss": 0.5009, |
| "step": 331 |
| }, |
| { |
| "epoch": 10.38, |
| "learning_rate": 0.00047549999999999996, |
| "loss": 0.5521, |
| "step": 332 |
| }, |
| { |
| "epoch": 10.41, |
| "learning_rate": 0.00047512499999999996, |
| "loss": 0.5354, |
| "step": 333 |
| }, |
| { |
| "epoch": 10.44, |
| "learning_rate": 0.00047474999999999997, |
| "loss": 0.4855, |
| "step": 334 |
| }, |
| { |
| "epoch": 10.47, |
| "learning_rate": 0.00047437499999999997, |
| "loss": 0.5356, |
| "step": 335 |
| }, |
| { |
| "epoch": 10.5, |
| "learning_rate": 0.000474, |
| "loss": 0.5868, |
| "step": 336 |
| }, |
| { |
| "epoch": 10.53, |
| "learning_rate": 0.000473625, |
| "loss": 0.5931, |
| "step": 337 |
| }, |
| { |
| "epoch": 10.56, |
| "learning_rate": 0.00047324999999999993, |
| "loss": 0.5786, |
| "step": 338 |
| }, |
| { |
| "epoch": 10.59, |
| "learning_rate": 0.00047287499999999993, |
| "loss": 0.5718, |
| "step": 339 |
| }, |
| { |
| "epoch": 10.62, |
| "learning_rate": 0.00047249999999999994, |
| "loss": 0.6153, |
| "step": 340 |
| }, |
| { |
| "epoch": 10.62, |
| "eval_loss": 0.9987176060676575, |
| "eval_mean_accuracy": 0.5497999939840363, |
| "eval_mean_iou": 0.3857005670859029, |
| "eval_overall_accuracy": 0.660497130009166, |
| "eval_per_category_accuracy": [ |
| 0.5231205889958428, |
| 0.7747069218397336, |
| 0.7498118814452185, |
| 0.3063129332898453, |
| 0.546321263677379, |
| 0.6574151027622224, |
| 0.6809347098082498, |
| 0.45994636300225755, |
| 0.4333802117068756, |
| 0.5861572233919434, |
| 0.342909526097829, |
| 0.5365832017910386 |
| ], |
| "eval_per_category_iou": [ |
| 0.4772630084705572, |
| 0.5541882497823591, |
| 0.5825233033041953, |
| 0.24382429625701532, |
| 0.38442721103951644, |
| 0.38358193468602964, |
| 0.42849690958511605, |
| 0.3960627429416036, |
| 0.21535871992663122, |
| 0.4152440328818222, |
| 0.1834708686588136, |
| 0.36396552749717453 |
| ], |
| "eval_runtime": 123.7757, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 340 |
| }, |
| { |
| "epoch": 10.66, |
| "learning_rate": 0.00047212499999999994, |
| "loss": 0.6764, |
| "step": 341 |
| }, |
| { |
| "epoch": 10.69, |
| "learning_rate": 0.00047174999999999995, |
| "loss": 0.541, |
| "step": 342 |
| }, |
| { |
| "epoch": 10.72, |
| "learning_rate": 0.00047137499999999995, |
| "loss": 0.5097, |
| "step": 343 |
| }, |
| { |
| "epoch": 10.75, |
| "learning_rate": 0.00047099999999999996, |
| "loss": 0.4856, |
| "step": 344 |
| }, |
| { |
| "epoch": 10.78, |
| "learning_rate": 0.00047062499999999996, |
| "loss": 0.5468, |
| "step": 345 |
| }, |
| { |
| "epoch": 10.81, |
| "learning_rate": 0.0004702499999999999, |
| "loss": 0.5513, |
| "step": 346 |
| }, |
| { |
| "epoch": 10.84, |
| "learning_rate": 0.0004698749999999999, |
| "loss": 0.6115, |
| "step": 347 |
| }, |
| { |
| "epoch": 10.88, |
| "learning_rate": 0.0004694999999999999, |
| "loss": 0.5045, |
| "step": 348 |
| }, |
| { |
| "epoch": 10.91, |
| "learning_rate": 0.0004691249999999999, |
| "loss": 0.5671, |
| "step": 349 |
| }, |
| { |
| "epoch": 10.94, |
| "learning_rate": 0.00046875, |
| "loss": 0.5062, |
| "step": 350 |
| }, |
| { |
| "epoch": 10.97, |
| "learning_rate": 0.000468375, |
| "loss": 0.6014, |
| "step": 351 |
| }, |
| { |
| "epoch": 11.0, |
| "learning_rate": 0.000468, |
| "loss": 0.5981, |
| "step": 352 |
| }, |
| { |
| "epoch": 11.03, |
| "learning_rate": 0.000467625, |
| "loss": 0.5369, |
| "step": 353 |
| }, |
| { |
| "epoch": 11.06, |
| "learning_rate": 0.00046725, |
| "loss": 0.4975, |
| "step": 354 |
| }, |
| { |
| "epoch": 11.09, |
| "learning_rate": 0.00046687499999999995, |
| "loss": 0.5137, |
| "step": 355 |
| }, |
| { |
| "epoch": 11.12, |
| "learning_rate": 0.00046649999999999996, |
| "loss": 0.4978, |
| "step": 356 |
| }, |
| { |
| "epoch": 11.16, |
| "learning_rate": 0.00046612499999999996, |
| "loss": 0.4579, |
| "step": 357 |
| }, |
| { |
| "epoch": 11.19, |
| "learning_rate": 0.00046574999999999996, |
| "loss": 0.5321, |
| "step": 358 |
| }, |
| { |
| "epoch": 11.22, |
| "learning_rate": 0.00046537499999999997, |
| "loss": 0.5201, |
| "step": 359 |
| }, |
| { |
| "epoch": 11.25, |
| "learning_rate": 0.00046499999999999997, |
| "loss": 0.4252, |
| "step": 360 |
| }, |
| { |
| "epoch": 11.25, |
| "eval_loss": 0.9291033148765564, |
| "eval_mean_accuracy": 0.528009973276733, |
| "eval_mean_iou": 0.39591471843488896, |
| "eval_overall_accuracy": 0.6944302341394257, |
| "eval_per_category_accuracy": [ |
| 0.7758503951264145, |
| 0.7639770669563318, |
| 0.6693553518837109, |
| 0.38660037057341723, |
| 0.5337667415575549, |
| 0.5057342154831077, |
| 0.5763875877943601, |
| 0.6287229015837165, |
| 0.3423444806260703, |
| 0.5546978372878139, |
| 0.15132853605458002, |
| 0.447354194393719 |
| ], |
| "eval_per_category_iou": [ |
| 0.6138349391900487, |
| 0.5748497846975844, |
| 0.5686854787211605, |
| 0.28537363740676996, |
| 0.3701880790533382, |
| 0.3687162885337687, |
| 0.4387556373721067, |
| 0.4152303639246888, |
| 0.22665730662981398, |
| 0.4070483981278673, |
| 0.12507399555773677, |
| 0.35656271200378264 |
| ], |
| "eval_runtime": 123.9322, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 360 |
| }, |
| { |
| "epoch": 11.28, |
| "learning_rate": 0.000464625, |
| "loss": 0.4748, |
| "step": 361 |
| }, |
| { |
| "epoch": 11.31, |
| "learning_rate": 0.00046425, |
| "loss": 0.5546, |
| "step": 362 |
| }, |
| { |
| "epoch": 11.34, |
| "learning_rate": 0.00046387499999999993, |
| "loss": 0.5495, |
| "step": 363 |
| }, |
| { |
| "epoch": 11.38, |
| "learning_rate": 0.00046349999999999994, |
| "loss": 0.5273, |
| "step": 364 |
| }, |
| { |
| "epoch": 11.41, |
| "learning_rate": 0.00046312499999999994, |
| "loss": 0.4722, |
| "step": 365 |
| }, |
| { |
| "epoch": 11.44, |
| "learning_rate": 0.00046274999999999995, |
| "loss": 0.535, |
| "step": 366 |
| }, |
| { |
| "epoch": 11.47, |
| "learning_rate": 0.00046237499999999995, |
| "loss": 0.5302, |
| "step": 367 |
| }, |
| { |
| "epoch": 11.5, |
| "learning_rate": 0.00046199999999999995, |
| "loss": 0.5074, |
| "step": 368 |
| }, |
| { |
| "epoch": 11.53, |
| "learning_rate": 0.00046162499999999996, |
| "loss": 0.5291, |
| "step": 369 |
| }, |
| { |
| "epoch": 11.56, |
| "learning_rate": 0.00046124999999999996, |
| "loss": 0.4986, |
| "step": 370 |
| }, |
| { |
| "epoch": 11.59, |
| "learning_rate": 0.0004608749999999999, |
| "loss": 0.4701, |
| "step": 371 |
| }, |
| { |
| "epoch": 11.62, |
| "learning_rate": 0.0004604999999999999, |
| "loss": 0.5317, |
| "step": 372 |
| }, |
| { |
| "epoch": 11.66, |
| "learning_rate": 0.0004601249999999999, |
| "loss": 0.4986, |
| "step": 373 |
| }, |
| { |
| "epoch": 11.69, |
| "learning_rate": 0.0004597499999999999, |
| "loss": 0.5165, |
| "step": 374 |
| }, |
| { |
| "epoch": 11.72, |
| "learning_rate": 0.000459375, |
| "loss": 0.5088, |
| "step": 375 |
| }, |
| { |
| "epoch": 11.75, |
| "learning_rate": 0.000459, |
| "loss": 0.4453, |
| "step": 376 |
| }, |
| { |
| "epoch": 11.78, |
| "learning_rate": 0.000458625, |
| "loss": 0.5066, |
| "step": 377 |
| }, |
| { |
| "epoch": 11.81, |
| "learning_rate": 0.00045825, |
| "loss": 0.6757, |
| "step": 378 |
| }, |
| { |
| "epoch": 11.84, |
| "learning_rate": 0.000457875, |
| "loss": 0.5308, |
| "step": 379 |
| }, |
| { |
| "epoch": 11.88, |
| "learning_rate": 0.00045749999999999995, |
| "loss": 0.511, |
| "step": 380 |
| }, |
| { |
| "epoch": 11.88, |
| "eval_loss": 1.032415509223938, |
| "eval_mean_accuracy": 0.5351733097568384, |
| "eval_mean_iou": 0.3874038254982633, |
| "eval_overall_accuracy": 0.6665094944468716, |
| "eval_per_category_accuracy": [ |
| 0.5126931341233072, |
| 0.8282813330437976, |
| 0.6890960293502041, |
| 0.5476887738461218, |
| 0.6291867821971731, |
| 0.4474030899437103, |
| 0.6554430535197253, |
| 0.4929760819276747, |
| 0.3958812099056813, |
| 0.5960181728949322, |
| 0.2651868765552641, |
| 0.36222517977446944 |
| ], |
| "eval_per_category_iou": [ |
| 0.4716454125121508, |
| 0.5641064988156196, |
| 0.574187815846751, |
| 0.2775623677866228, |
| 0.39063569128597303, |
| 0.3542157527555254, |
| 0.4394969112521424, |
| 0.40979043774810314, |
| 0.23025514522221993, |
| 0.42890476310219977, |
| 0.19330655957161982, |
| 0.3147385500802324 |
| ], |
| "eval_runtime": 123.8124, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 380 |
| }, |
| { |
| "epoch": 11.91, |
| "learning_rate": 0.00045712499999999996, |
| "loss": 0.5862, |
| "step": 381 |
| }, |
| { |
| "epoch": 11.94, |
| "learning_rate": 0.00045674999999999996, |
| "loss": 0.5058, |
| "step": 382 |
| }, |
| { |
| "epoch": 11.97, |
| "learning_rate": 0.00045637499999999997, |
| "loss": 0.511, |
| "step": 383 |
| }, |
| { |
| "epoch": 12.0, |
| "learning_rate": 0.00045599999999999997, |
| "loss": 0.5438, |
| "step": 384 |
| }, |
| { |
| "epoch": 12.03, |
| "learning_rate": 0.000455625, |
| "loss": 0.447, |
| "step": 385 |
| }, |
| { |
| "epoch": 12.06, |
| "learning_rate": 0.00045525, |
| "loss": 0.4158, |
| "step": 386 |
| }, |
| { |
| "epoch": 12.09, |
| "learning_rate": 0.000454875, |
| "loss": 0.4493, |
| "step": 387 |
| }, |
| { |
| "epoch": 12.12, |
| "learning_rate": 0.00045449999999999993, |
| "loss": 0.515, |
| "step": 388 |
| }, |
| { |
| "epoch": 12.16, |
| "learning_rate": 0.00045412499999999994, |
| "loss": 0.5064, |
| "step": 389 |
| }, |
| { |
| "epoch": 12.19, |
| "learning_rate": 0.00045374999999999994, |
| "loss": 0.5294, |
| "step": 390 |
| }, |
| { |
| "epoch": 12.22, |
| "learning_rate": 0.00045337499999999995, |
| "loss": 0.4611, |
| "step": 391 |
| }, |
| { |
| "epoch": 12.25, |
| "learning_rate": 0.00045299999999999995, |
| "loss": 0.5199, |
| "step": 392 |
| }, |
| { |
| "epoch": 12.28, |
| "learning_rate": 0.00045262499999999996, |
| "loss": 0.494, |
| "step": 393 |
| }, |
| { |
| "epoch": 12.31, |
| "learning_rate": 0.00045224999999999996, |
| "loss": 0.5388, |
| "step": 394 |
| }, |
| { |
| "epoch": 12.34, |
| "learning_rate": 0.00045187499999999997, |
| "loss": 0.4428, |
| "step": 395 |
| }, |
| { |
| "epoch": 12.38, |
| "learning_rate": 0.0004514999999999999, |
| "loss": 0.5291, |
| "step": 396 |
| }, |
| { |
| "epoch": 12.41, |
| "learning_rate": 0.0004511249999999999, |
| "loss": 0.5459, |
| "step": 397 |
| }, |
| { |
| "epoch": 12.44, |
| "learning_rate": 0.0004507499999999999, |
| "loss": 0.5026, |
| "step": 398 |
| }, |
| { |
| "epoch": 12.47, |
| "learning_rate": 0.00045037499999999993, |
| "loss": 0.5689, |
| "step": 399 |
| }, |
| { |
| "epoch": 12.5, |
| "learning_rate": 0.00045, |
| "loss": 0.4396, |
| "step": 400 |
| }, |
| { |
| "epoch": 12.5, |
| "eval_loss": 0.9480527639389038, |
| "eval_mean_accuracy": 0.530182694199835, |
| "eval_mean_iou": 0.4007987151571945, |
| "eval_overall_accuracy": 0.6919526886521724, |
| "eval_per_category_accuracy": [ |
| 0.6883764832199943, |
| 0.7711528595241125, |
| 0.728660136472628, |
| 0.2644395250553431, |
| 0.4997416973422176, |
| 0.5449627820381144, |
| 0.6418612010725042, |
| 0.4704519715293187, |
| 0.30828593737522514, |
| 0.6739508105150778, |
| 0.24415006442963702, |
| 0.5261588618238469 |
| ], |
| "eval_per_category_iou": [ |
| 0.5870857495659079, |
| 0.5727333442653406, |
| 0.5911691283424428, |
| 0.23602646053218374, |
| 0.3928900215019511, |
| 0.3900904704071493, |
| 0.44108434621351955, |
| 0.40724866601060883, |
| 0.22723356109173612, |
| 0.4344269928246981, |
| 0.17749152783742347, |
| 0.35210431329337216 |
| ], |
| "eval_runtime": 124.6159, |
| "eval_samples_per_second": 0.915, |
| "eval_steps_per_second": 0.08, |
| "step": 400 |
| }, |
| { |
| "epoch": 12.53, |
| "learning_rate": 0.000449625, |
| "loss": 0.5228, |
| "step": 401 |
| }, |
| { |
| "epoch": 12.56, |
| "learning_rate": 0.00044925, |
| "loss": 0.4921, |
| "step": 402 |
| }, |
| { |
| "epoch": 12.59, |
| "learning_rate": 0.000448875, |
| "loss": 0.5139, |
| "step": 403 |
| }, |
| { |
| "epoch": 12.62, |
| "learning_rate": 0.0004485, |
| "loss": 0.5392, |
| "step": 404 |
| }, |
| { |
| "epoch": 12.66, |
| "learning_rate": 0.00044812499999999996, |
| "loss": 0.5223, |
| "step": 405 |
| }, |
| { |
| "epoch": 12.69, |
| "learning_rate": 0.00044774999999999996, |
| "loss": 0.4715, |
| "step": 406 |
| }, |
| { |
| "epoch": 12.72, |
| "learning_rate": 0.00044737499999999996, |
| "loss": 0.4516, |
| "step": 407 |
| }, |
| { |
| "epoch": 12.75, |
| "learning_rate": 0.00044699999999999997, |
| "loss": 0.5014, |
| "step": 408 |
| }, |
| { |
| "epoch": 12.78, |
| "learning_rate": 0.000446625, |
| "loss": 0.4476, |
| "step": 409 |
| }, |
| { |
| "epoch": 12.81, |
| "learning_rate": 0.00044625, |
| "loss": 0.5565, |
| "step": 410 |
| }, |
| { |
| "epoch": 12.84, |
| "learning_rate": 0.000445875, |
| "loss": 0.431, |
| "step": 411 |
| }, |
| { |
| "epoch": 12.88, |
| "learning_rate": 0.0004455, |
| "loss": 0.4688, |
| "step": 412 |
| }, |
| { |
| "epoch": 12.91, |
| "learning_rate": 0.00044512499999999994, |
| "loss": 0.4657, |
| "step": 413 |
| }, |
| { |
| "epoch": 12.94, |
| "learning_rate": 0.00044474999999999994, |
| "loss": 0.4364, |
| "step": 414 |
| }, |
| { |
| "epoch": 12.97, |
| "learning_rate": 0.00044437499999999995, |
| "loss": 0.5438, |
| "step": 415 |
| }, |
| { |
| "epoch": 13.0, |
| "learning_rate": 0.00044399999999999995, |
| "loss": 0.4812, |
| "step": 416 |
| }, |
| { |
| "epoch": 13.03, |
| "learning_rate": 0.00044362499999999995, |
| "loss": 0.4611, |
| "step": 417 |
| }, |
| { |
| "epoch": 13.06, |
| "learning_rate": 0.00044324999999999996, |
| "loss": 0.5335, |
| "step": 418 |
| }, |
| { |
| "epoch": 13.09, |
| "learning_rate": 0.00044287499999999996, |
| "loss": 0.4892, |
| "step": 419 |
| }, |
| { |
| "epoch": 13.12, |
| "learning_rate": 0.00044249999999999997, |
| "loss": 0.433, |
| "step": 420 |
| }, |
| { |
| "epoch": 13.12, |
| "eval_loss": 0.9964607357978821, |
| "eval_mean_accuracy": 0.5161420143998455, |
| "eval_mean_iou": 0.38527011433114106, |
| "eval_overall_accuracy": 0.6863714519299959, |
| "eval_per_category_accuracy": [ |
| 0.6875807532313407, |
| 0.7827795368926075, |
| 0.7752045334502897, |
| 0.3332280830806171, |
| 0.6272763554811824, |
| 0.5677653617092318, |
| 0.5813202886251887, |
| 0.35458598473606334, |
| 0.3958146633186339, |
| 0.5571871963599969, |
| 0.18530024823009797, |
| 0.3456611676828954 |
| ], |
| "eval_per_category_iou": [ |
| 0.5768863441650554, |
| 0.5722268835575453, |
| 0.5735459134286014, |
| 0.2657191158282677, |
| 0.3714637057832005, |
| 0.36598911915704463, |
| 0.44648066580609275, |
| 0.3262031822008087, |
| 0.2515546119567317, |
| 0.42096691698866495, |
| 0.15479219135526265, |
| 0.29741272174641725 |
| ], |
| "eval_runtime": 123.9379, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 420 |
| }, |
| { |
| "epoch": 13.16, |
| "learning_rate": 0.0004421249999999999, |
| "loss": 0.5184, |
| "step": 421 |
| }, |
| { |
| "epoch": 13.19, |
| "learning_rate": 0.0004417499999999999, |
| "loss": 0.4831, |
| "step": 422 |
| }, |
| { |
| "epoch": 13.22, |
| "learning_rate": 0.00044137499999999993, |
| "loss": 0.5038, |
| "step": 423 |
| }, |
| { |
| "epoch": 13.25, |
| "learning_rate": 0.00044099999999999993, |
| "loss": 0.5039, |
| "step": 424 |
| }, |
| { |
| "epoch": 13.28, |
| "learning_rate": 0.00044062499999999994, |
| "loss": 0.55, |
| "step": 425 |
| }, |
| { |
| "epoch": 13.31, |
| "learning_rate": 0.00044025, |
| "loss": 0.4896, |
| "step": 426 |
| }, |
| { |
| "epoch": 13.34, |
| "learning_rate": 0.000439875, |
| "loss": 0.5348, |
| "step": 427 |
| }, |
| { |
| "epoch": 13.38, |
| "learning_rate": 0.0004395, |
| "loss": 0.5004, |
| "step": 428 |
| }, |
| { |
| "epoch": 13.41, |
| "learning_rate": 0.000439125, |
| "loss": 0.448, |
| "step": 429 |
| }, |
| { |
| "epoch": 13.44, |
| "learning_rate": 0.00043874999999999996, |
| "loss": 0.4679, |
| "step": 430 |
| }, |
| { |
| "epoch": 13.47, |
| "learning_rate": 0.00043837499999999996, |
| "loss": 0.5242, |
| "step": 431 |
| }, |
| { |
| "epoch": 13.5, |
| "learning_rate": 0.00043799999999999997, |
| "loss": 0.4008, |
| "step": 432 |
| }, |
| { |
| "epoch": 13.53, |
| "learning_rate": 0.00043762499999999997, |
| "loss": 0.4851, |
| "step": 433 |
| }, |
| { |
| "epoch": 13.56, |
| "learning_rate": 0.00043725, |
| "loss": 0.4287, |
| "step": 434 |
| }, |
| { |
| "epoch": 13.59, |
| "learning_rate": 0.000436875, |
| "loss": 0.4648, |
| "step": 435 |
| }, |
| { |
| "epoch": 13.62, |
| "learning_rate": 0.0004365, |
| "loss": 0.4869, |
| "step": 436 |
| }, |
| { |
| "epoch": 13.66, |
| "learning_rate": 0.000436125, |
| "loss": 0.5123, |
| "step": 437 |
| }, |
| { |
| "epoch": 13.69, |
| "learning_rate": 0.00043574999999999994, |
| "loss": 0.4866, |
| "step": 438 |
| }, |
| { |
| "epoch": 13.72, |
| "learning_rate": 0.00043537499999999994, |
| "loss": 0.473, |
| "step": 439 |
| }, |
| { |
| "epoch": 13.75, |
| "learning_rate": 0.00043499999999999995, |
| "loss": 0.391, |
| "step": 440 |
| }, |
| { |
| "epoch": 13.75, |
| "eval_loss": 0.960097074508667, |
| "eval_mean_accuracy": 0.5366096582570204, |
| "eval_mean_iou": 0.3954851774486026, |
| "eval_overall_accuracy": 0.6906181469298246, |
| "eval_per_category_accuracy": [ |
| 0.6962806267574084, |
| 0.7853961112845703, |
| 0.7287651461475003, |
| 0.45656865670606034, |
| 0.5133213264803781, |
| 0.5403862974025455, |
| 0.6444582294863095, |
| 0.5842254082508402, |
| 0.3959765933471159, |
| 0.5184829565220113, |
| 0.23149981788297375, |
| 0.34395472881653205 |
| ], |
| "eval_per_category_iou": [ |
| 0.5852621382518784, |
| 0.5768371325534494, |
| 0.5832121882254719, |
| 0.2472658393403438, |
| 0.3798774216569066, |
| 0.37300094191208155, |
| 0.45491619947198525, |
| 0.4265556855029551, |
| 0.23847794983053747, |
| 0.4091629174845173, |
| 0.1723333378903669, |
| 0.29892037726273807 |
| ], |
| "eval_runtime": 123.423, |
| "eval_samples_per_second": 0.924, |
| "eval_steps_per_second": 0.081, |
| "step": 440 |
| }, |
| { |
| "epoch": 13.78, |
| "learning_rate": 0.00043462499999999995, |
| "loss": 0.4544, |
| "step": 441 |
| }, |
| { |
| "epoch": 13.81, |
| "learning_rate": 0.00043424999999999996, |
| "loss": 0.4717, |
| "step": 442 |
| }, |
| { |
| "epoch": 13.84, |
| "learning_rate": 0.00043387499999999996, |
| "loss": 0.4542, |
| "step": 443 |
| }, |
| { |
| "epoch": 13.88, |
| "learning_rate": 0.00043349999999999997, |
| "loss": 0.444, |
| "step": 444 |
| }, |
| { |
| "epoch": 13.91, |
| "learning_rate": 0.00043312499999999997, |
| "loss": 0.4853, |
| "step": 445 |
| }, |
| { |
| "epoch": 13.94, |
| "learning_rate": 0.0004327499999999999, |
| "loss": 0.499, |
| "step": 446 |
| }, |
| { |
| "epoch": 13.97, |
| "learning_rate": 0.0004323749999999999, |
| "loss": 0.479, |
| "step": 447 |
| }, |
| { |
| "epoch": 14.0, |
| "learning_rate": 0.00043199999999999993, |
| "loss": 0.4411, |
| "step": 448 |
| }, |
| { |
| "epoch": 14.03, |
| "learning_rate": 0.00043162499999999993, |
| "loss": 0.4395, |
| "step": 449 |
| }, |
| { |
| "epoch": 14.06, |
| "learning_rate": 0.00043124999999999994, |
| "loss": 0.4041, |
| "step": 450 |
| }, |
| { |
| "epoch": 14.09, |
| "learning_rate": 0.000430875, |
| "loss": 0.4702, |
| "step": 451 |
| }, |
| { |
| "epoch": 14.12, |
| "learning_rate": 0.0004305, |
| "loss": 0.4038, |
| "step": 452 |
| }, |
| { |
| "epoch": 14.16, |
| "learning_rate": 0.000430125, |
| "loss": 0.4225, |
| "step": 453 |
| }, |
| { |
| "epoch": 14.19, |
| "learning_rate": 0.00042975, |
| "loss": 0.4695, |
| "step": 454 |
| }, |
| { |
| "epoch": 14.22, |
| "learning_rate": 0.00042937499999999996, |
| "loss": 0.4831, |
| "step": 455 |
| }, |
| { |
| "epoch": 14.25, |
| "learning_rate": 0.00042899999999999997, |
| "loss": 0.4788, |
| "step": 456 |
| }, |
| { |
| "epoch": 14.28, |
| "learning_rate": 0.00042862499999999997, |
| "loss": 0.4588, |
| "step": 457 |
| }, |
| { |
| "epoch": 14.31, |
| "learning_rate": 0.00042825, |
| "loss": 0.4695, |
| "step": 458 |
| }, |
| { |
| "epoch": 14.34, |
| "learning_rate": 0.000427875, |
| "loss": 0.4397, |
| "step": 459 |
| }, |
| { |
| "epoch": 14.38, |
| "learning_rate": 0.0004275, |
| "loss": 0.3838, |
| "step": 460 |
| }, |
| { |
| "epoch": 14.38, |
| "eval_loss": 1.0273075103759766, |
| "eval_mean_accuracy": 0.5082167691158227, |
| "eval_mean_iou": 0.3857975412859527, |
| "eval_overall_accuracy": 0.6752951103344298, |
| "eval_per_category_accuracy": [ |
| 0.5440038390043963, |
| 0.8360225611529776, |
| 0.758805046970428, |
| 0.3064343758891333, |
| 0.6123616561353465, |
| 0.5102689637401602, |
| 0.5372959730708857, |
| 0.4196982570654369, |
| 0.40250924997559956, |
| 0.5669620900811063, |
| 0.19339909951547687, |
| 0.4108401167889236 |
| ], |
| "eval_per_category_iou": [ |
| 0.49498153666173117, |
| 0.5712528393596524, |
| 0.5812824830895571, |
| 0.25617559696470427, |
| 0.38368903348485445, |
| 0.38378641015212356, |
| 0.43791696768969585, |
| 0.3702304506367591, |
| 0.25293454306714636, |
| 0.4225529002066759, |
| 0.15632514918504883, |
| 0.31844258493348304 |
| ], |
| "eval_runtime": 124.352, |
| "eval_samples_per_second": 0.917, |
| "eval_steps_per_second": 0.08, |
| "step": 460 |
| }, |
| { |
| "epoch": 14.41, |
| "learning_rate": 0.000427125, |
| "loss": 0.4693, |
| "step": 461 |
| }, |
| { |
| "epoch": 14.44, |
| "learning_rate": 0.00042675, |
| "loss": 0.4268, |
| "step": 462 |
| }, |
| { |
| "epoch": 14.47, |
| "learning_rate": 0.00042637499999999994, |
| "loss": 0.4533, |
| "step": 463 |
| }, |
| { |
| "epoch": 14.5, |
| "learning_rate": 0.00042599999999999995, |
| "loss": 0.4642, |
| "step": 464 |
| }, |
| { |
| "epoch": 14.53, |
| "learning_rate": 0.00042562499999999995, |
| "loss": 0.4769, |
| "step": 465 |
| }, |
| { |
| "epoch": 14.56, |
| "learning_rate": 0.00042524999999999996, |
| "loss": 0.4274, |
| "step": 466 |
| }, |
| { |
| "epoch": 14.59, |
| "learning_rate": 0.00042487499999999996, |
| "loss": 0.4775, |
| "step": 467 |
| }, |
| { |
| "epoch": 14.62, |
| "learning_rate": 0.00042449999999999996, |
| "loss": 0.3955, |
| "step": 468 |
| }, |
| { |
| "epoch": 14.66, |
| "learning_rate": 0.00042412499999999997, |
| "loss": 0.4803, |
| "step": 469 |
| }, |
| { |
| "epoch": 14.69, |
| "learning_rate": 0.00042375, |
| "loss": 0.3911, |
| "step": 470 |
| }, |
| { |
| "epoch": 14.72, |
| "learning_rate": 0.0004233749999999999, |
| "loss": 0.4764, |
| "step": 471 |
| }, |
| { |
| "epoch": 14.75, |
| "learning_rate": 0.00042299999999999993, |
| "loss": 0.4527, |
| "step": 472 |
| }, |
| { |
| "epoch": 14.78, |
| "learning_rate": 0.00042262499999999993, |
| "loss": 0.4481, |
| "step": 473 |
| }, |
| { |
| "epoch": 14.81, |
| "learning_rate": 0.00042224999999999994, |
| "loss": 0.5046, |
| "step": 474 |
| }, |
| { |
| "epoch": 14.84, |
| "learning_rate": 0.00042187499999999994, |
| "loss": 0.5169, |
| "step": 475 |
| }, |
| { |
| "epoch": 14.88, |
| "learning_rate": 0.00042149999999999995, |
| "loss": 0.4903, |
| "step": 476 |
| }, |
| { |
| "epoch": 14.91, |
| "learning_rate": 0.000421125, |
| "loss": 0.4304, |
| "step": 477 |
| }, |
| { |
| "epoch": 14.94, |
| "learning_rate": 0.00042075, |
| "loss": 0.4314, |
| "step": 478 |
| }, |
| { |
| "epoch": 14.97, |
| "learning_rate": 0.000420375, |
| "loss": 0.4919, |
| "step": 479 |
| }, |
| { |
| "epoch": 15.0, |
| "learning_rate": 0.00041999999999999996, |
| "loss": 0.439, |
| "step": 480 |
| }, |
| { |
| "epoch": 15.0, |
| "eval_loss": 0.9900245070457458, |
| "eval_mean_accuracy": 0.5459215103584506, |
| "eval_mean_iou": 0.3977724156715366, |
| "eval_overall_accuracy": 0.688731076424582, |
| "eval_per_category_accuracy": [ |
| 0.6912638127580818, |
| 0.7736361191880559, |
| 0.7429642804454579, |
| 0.44192614902047866, |
| 0.5566209555375025, |
| 0.5489132209959389, |
| 0.6376629735699813, |
| 0.5306998417462132, |
| 0.4381382926807627, |
| 0.5487234798246131, |
| 0.27793200760300324, |
| 0.36257699093131945 |
| ], |
| "eval_per_category_iou": [ |
| 0.5862413617170502, |
| 0.5748060664442914, |
| 0.5872486090000612, |
| 0.2653964046751309, |
| 0.38542480588619005, |
| 0.38432922103497785, |
| 0.4162172327752406, |
| 0.4191820285272756, |
| 0.24282352825470666, |
| 0.41710927412113163, |
| 0.18954738162682103, |
| 0.30494307399556253 |
| ], |
| "eval_runtime": 123.2936, |
| "eval_samples_per_second": 0.925, |
| "eval_steps_per_second": 0.081, |
| "step": 480 |
| }, |
| { |
| "epoch": 15.03, |
| "learning_rate": 0.00041962499999999997, |
| "loss": 0.4175, |
| "step": 481 |
| }, |
| { |
| "epoch": 15.06, |
| "learning_rate": 0.00041924999999999997, |
| "loss": 0.4352, |
| "step": 482 |
| }, |
| { |
| "epoch": 15.09, |
| "learning_rate": 0.000418875, |
| "loss": 0.4637, |
| "step": 483 |
| }, |
| { |
| "epoch": 15.12, |
| "learning_rate": 0.0004185, |
| "loss": 0.45, |
| "step": 484 |
| }, |
| { |
| "epoch": 15.16, |
| "learning_rate": 0.000418125, |
| "loss": 0.3973, |
| "step": 485 |
| }, |
| { |
| "epoch": 15.19, |
| "learning_rate": 0.00041775, |
| "loss": 0.3689, |
| "step": 486 |
| }, |
| { |
| "epoch": 15.22, |
| "learning_rate": 0.000417375, |
| "loss": 0.4343, |
| "step": 487 |
| }, |
| { |
| "epoch": 15.25, |
| "learning_rate": 0.00041699999999999994, |
| "loss": 0.4032, |
| "step": 488 |
| }, |
| { |
| "epoch": 15.28, |
| "learning_rate": 0.00041662499999999995, |
| "loss": 0.407, |
| "step": 489 |
| }, |
| { |
| "epoch": 15.31, |
| "learning_rate": 0.00041624999999999995, |
| "loss": 0.4533, |
| "step": 490 |
| }, |
| { |
| "epoch": 15.34, |
| "learning_rate": 0.00041587499999999996, |
| "loss": 0.4727, |
| "step": 491 |
| }, |
| { |
| "epoch": 15.38, |
| "learning_rate": 0.00041549999999999996, |
| "loss": 0.4107, |
| "step": 492 |
| }, |
| { |
| "epoch": 15.41, |
| "learning_rate": 0.00041512499999999997, |
| "loss": 0.3918, |
| "step": 493 |
| }, |
| { |
| "epoch": 15.44, |
| "learning_rate": 0.00041474999999999997, |
| "loss": 0.4307, |
| "step": 494 |
| }, |
| { |
| "epoch": 15.47, |
| "learning_rate": 0.000414375, |
| "loss": 0.4523, |
| "step": 495 |
| }, |
| { |
| "epoch": 15.5, |
| "learning_rate": 0.0004139999999999999, |
| "loss": 0.412, |
| "step": 496 |
| }, |
| { |
| "epoch": 15.53, |
| "learning_rate": 0.00041362499999999993, |
| "loss": 0.4073, |
| "step": 497 |
| }, |
| { |
| "epoch": 15.56, |
| "learning_rate": 0.00041324999999999993, |
| "loss": 0.4751, |
| "step": 498 |
| }, |
| { |
| "epoch": 15.59, |
| "learning_rate": 0.00041287499999999994, |
| "loss": 0.3989, |
| "step": 499 |
| }, |
| { |
| "epoch": 15.62, |
| "learning_rate": 0.00041249999999999994, |
| "loss": 0.4512, |
| "step": 500 |
| }, |
| { |
| "epoch": 15.62, |
| "eval_loss": 0.9641626477241516, |
| "eval_mean_accuracy": 0.5429231404512459, |
| "eval_mean_iou": 0.4048638129614841, |
| "eval_overall_accuracy": 0.7011385800545675, |
| "eval_per_category_accuracy": [ |
| 0.795635716896333, |
| 0.7399733014822432, |
| 0.7527407382465496, |
| 0.30102150574943964, |
| 0.5432094763648003, |
| 0.681368154780267, |
| 0.5919154403140051, |
| 0.4270366469483377, |
| 0.3322671091275299, |
| 0.5841200483372719, |
| 0.23177528901512948, |
| 0.534014258153043 |
| ], |
| "eval_per_category_iou": [ |
| 0.6288522197434481, |
| 0.5748730843314954, |
| 0.5946081799214863, |
| 0.26243983894534, |
| 0.3784144618645802, |
| 0.3748662215371345, |
| 0.45285274542078546, |
| 0.3886880444863579, |
| 0.22822951853764895, |
| 0.4246046233899173, |
| 0.18069362451500212, |
| 0.36924319284461304 |
| ], |
| "eval_runtime": 125.0285, |
| "eval_samples_per_second": 0.912, |
| "eval_steps_per_second": 0.08, |
| "step": 500 |
| }, |
| { |
| "epoch": 15.66, |
| "learning_rate": 0.00041212499999999995, |
| "loss": 0.3799, |
| "step": 501 |
| }, |
| { |
| "epoch": 15.69, |
| "learning_rate": 0.00041175, |
| "loss": 0.4713, |
| "step": 502 |
| }, |
| { |
| "epoch": 15.72, |
| "learning_rate": 0.000411375, |
| "loss": 0.396, |
| "step": 503 |
| }, |
| { |
| "epoch": 15.75, |
| "learning_rate": 0.000411, |
| "loss": 0.4217, |
| "step": 504 |
| }, |
| { |
| "epoch": 15.78, |
| "learning_rate": 0.0004106249999999999, |
| "loss": 0.4644, |
| "step": 505 |
| }, |
| { |
| "epoch": 15.81, |
| "learning_rate": 0.00041024999999999997, |
| "loss": 0.4954, |
| "step": 506 |
| }, |
| { |
| "epoch": 15.84, |
| "learning_rate": 0.000409875, |
| "loss": 0.5335, |
| "step": 507 |
| }, |
| { |
| "epoch": 15.88, |
| "learning_rate": 0.0004095, |
| "loss": 0.4492, |
| "step": 508 |
| }, |
| { |
| "epoch": 15.91, |
| "learning_rate": 0.000409125, |
| "loss": 0.4213, |
| "step": 509 |
| }, |
| { |
| "epoch": 15.94, |
| "learning_rate": 0.00040875, |
| "loss": 0.3764, |
| "step": 510 |
| }, |
| { |
| "epoch": 15.97, |
| "learning_rate": 0.000408375, |
| "loss": 0.4786, |
| "step": 511 |
| }, |
| { |
| "epoch": 16.0, |
| "learning_rate": 0.000408, |
| "loss": 0.4795, |
| "step": 512 |
| }, |
| { |
| "epoch": 16.03, |
| "learning_rate": 0.00040762499999999995, |
| "loss": 0.3959, |
| "step": 513 |
| }, |
| { |
| "epoch": 16.06, |
| "learning_rate": 0.00040724999999999995, |
| "loss": 0.4325, |
| "step": 514 |
| }, |
| { |
| "epoch": 16.09, |
| "learning_rate": 0.00040687499999999996, |
| "loss": 0.4145, |
| "step": 515 |
| }, |
| { |
| "epoch": 16.12, |
| "learning_rate": 0.00040649999999999996, |
| "loss": 0.397, |
| "step": 516 |
| }, |
| { |
| "epoch": 16.16, |
| "learning_rate": 0.00040612499999999996, |
| "loss": 0.4268, |
| "step": 517 |
| }, |
| { |
| "epoch": 16.19, |
| "learning_rate": 0.00040574999999999997, |
| "loss": 0.3969, |
| "step": 518 |
| }, |
| { |
| "epoch": 16.22, |
| "learning_rate": 0.000405375, |
| "loss": 0.4402, |
| "step": 519 |
| }, |
| { |
| "epoch": 16.25, |
| "learning_rate": 0.000405, |
| "loss": 0.4993, |
| "step": 520 |
| }, |
| { |
| "epoch": 16.25, |
| "eval_loss": 1.0489466190338135, |
| "eval_mean_accuracy": 0.5290564444706155, |
| "eval_mean_iou": 0.3977378405073158, |
| "eval_overall_accuracy": 0.683437514723393, |
| "eval_per_category_accuracy": [ |
| 0.5979805873600369, |
| 0.8259897587771458, |
| 0.6962783486886204, |
| 0.41443154454167563, |
| 0.46842731160619683, |
| 0.5239240180155985, |
| 0.5468496368568635, |
| 0.58211357200333, |
| 0.3897500510190501, |
| 0.6099821577678903, |
| 0.3102417106145149, |
| 0.3827086363964633 |
| ], |
| "eval_per_category_iou": [ |
| 0.5351354597494672, |
| 0.5752502813654251, |
| 0.5795456245827204, |
| 0.2539137237561544, |
| 0.3778761249961186, |
| 0.38376103706913395, |
| 0.4377072638680518, |
| 0.41509028884024585, |
| 0.24749133729610953, |
| 0.43313183205436867, |
| 0.20596560601965777, |
| 0.3279855064903372 |
| ], |
| "eval_runtime": 124.56, |
| "eval_samples_per_second": 0.915, |
| "eval_steps_per_second": 0.08, |
| "step": 520 |
| }, |
| { |
| "epoch": 16.28, |
| "learning_rate": 0.00040462499999999993, |
| "loss": 0.3526, |
| "step": 521 |
| }, |
| { |
| "epoch": 16.31, |
| "learning_rate": 0.00040424999999999993, |
| "loss": 0.4272, |
| "step": 522 |
| }, |
| { |
| "epoch": 16.34, |
| "learning_rate": 0.00040387499999999994, |
| "loss": 0.3533, |
| "step": 523 |
| }, |
| { |
| "epoch": 16.38, |
| "learning_rate": 0.00040349999999999994, |
| "loss": 0.3708, |
| "step": 524 |
| }, |
| { |
| "epoch": 16.41, |
| "learning_rate": 0.00040312499999999995, |
| "loss": 0.5067, |
| "step": 525 |
| }, |
| { |
| "epoch": 16.44, |
| "learning_rate": 0.00040274999999999995, |
| "loss": 0.4486, |
| "step": 526 |
| }, |
| { |
| "epoch": 16.47, |
| "learning_rate": 0.00040237499999999995, |
| "loss": 0.365, |
| "step": 527 |
| }, |
| { |
| "epoch": 16.5, |
| "learning_rate": 0.000402, |
| "loss": 0.3999, |
| "step": 528 |
| }, |
| { |
| "epoch": 16.53, |
| "learning_rate": 0.000401625, |
| "loss": 0.4774, |
| "step": 529 |
| }, |
| { |
| "epoch": 16.56, |
| "learning_rate": 0.0004012499999999999, |
| "loss": 0.4046, |
| "step": 530 |
| }, |
| { |
| "epoch": 16.59, |
| "learning_rate": 0.00040087499999999997, |
| "loss": 0.433, |
| "step": 531 |
| }, |
| { |
| "epoch": 16.62, |
| "learning_rate": 0.0004005, |
| "loss": 0.4509, |
| "step": 532 |
| }, |
| { |
| "epoch": 16.66, |
| "learning_rate": 0.000400125, |
| "loss": 0.4426, |
| "step": 533 |
| }, |
| { |
| "epoch": 16.69, |
| "learning_rate": 0.00039975, |
| "loss": 0.4442, |
| "step": 534 |
| }, |
| { |
| "epoch": 16.72, |
| "learning_rate": 0.000399375, |
| "loss": 0.463, |
| "step": 535 |
| }, |
| { |
| "epoch": 16.75, |
| "learning_rate": 0.000399, |
| "loss": 0.4252, |
| "step": 536 |
| }, |
| { |
| "epoch": 16.78, |
| "learning_rate": 0.000398625, |
| "loss": 0.3977, |
| "step": 537 |
| }, |
| { |
| "epoch": 16.81, |
| "learning_rate": 0.00039824999999999995, |
| "loss": 0.3889, |
| "step": 538 |
| }, |
| { |
| "epoch": 16.84, |
| "learning_rate": 0.00039787499999999995, |
| "loss": 0.4213, |
| "step": 539 |
| }, |
| { |
| "epoch": 16.88, |
| "learning_rate": 0.00039749999999999996, |
| "loss": 0.4365, |
| "step": 540 |
| }, |
| { |
| "epoch": 16.88, |
| "eval_loss": 0.9909974336624146, |
| "eval_mean_accuracy": 0.5328258993053784, |
| "eval_mean_iou": 0.4047647398185916, |
| "eval_overall_accuracy": 0.69754369635331, |
| "eval_per_category_accuracy": [ |
| 0.7311351861404658, |
| 0.7874778098484093, |
| 0.7424012802049058, |
| 0.3836788086134031, |
| 0.5249803993865565, |
| 0.5146730589862052, |
| 0.6326751675591081, |
| 0.46836906454547306, |
| 0.42196747202825124, |
| 0.4988820571608581, |
| 0.2849044880369009, |
| 0.4027659991540025 |
| ], |
| "eval_per_category_iou": [ |
| 0.6057512942967, |
| 0.5792283141071735, |
| 0.5886817077629869, |
| 0.2690359405564801, |
| 0.39189514134956965, |
| 0.3782952870642685, |
| 0.46891084995875004, |
| 0.40392528286893087, |
| 0.25226936068015177, |
| 0.3927578500285575, |
| 0.19890208748683166, |
| 0.32752376166269975 |
| ], |
| "eval_runtime": 123.418, |
| "eval_samples_per_second": 0.924, |
| "eval_steps_per_second": 0.081, |
| "step": 540 |
| }, |
| { |
| "epoch": 16.91, |
| "learning_rate": 0.00039712499999999996, |
| "loss": 0.4556, |
| "step": 541 |
| }, |
| { |
| "epoch": 16.94, |
| "learning_rate": 0.00039674999999999997, |
| "loss": 0.3967, |
| "step": 542 |
| }, |
| { |
| "epoch": 16.97, |
| "learning_rate": 0.00039637499999999997, |
| "loss": 0.4523, |
| "step": 543 |
| }, |
| { |
| "epoch": 17.0, |
| "learning_rate": 0.000396, |
| "loss": 0.2985, |
| "step": 544 |
| }, |
| { |
| "epoch": 17.03, |
| "learning_rate": 0.000395625, |
| "loss": 0.3761, |
| "step": 545 |
| }, |
| { |
| "epoch": 17.06, |
| "learning_rate": 0.00039524999999999993, |
| "loss": 0.3729, |
| "step": 546 |
| }, |
| { |
| "epoch": 17.09, |
| "learning_rate": 0.00039487499999999994, |
| "loss": 0.364, |
| "step": 547 |
| }, |
| { |
| "epoch": 17.12, |
| "learning_rate": 0.00039449999999999994, |
| "loss": 0.3819, |
| "step": 548 |
| }, |
| { |
| "epoch": 17.16, |
| "learning_rate": 0.00039412499999999994, |
| "loss": 0.4212, |
| "step": 549 |
| }, |
| { |
| "epoch": 17.19, |
| "learning_rate": 0.00039374999999999995, |
| "loss": 0.4293, |
| "step": 550 |
| }, |
| { |
| "epoch": 17.22, |
| "learning_rate": 0.00039337499999999995, |
| "loss": 0.3485, |
| "step": 551 |
| }, |
| { |
| "epoch": 17.25, |
| "learning_rate": 0.00039299999999999996, |
| "loss": 0.4189, |
| "step": 552 |
| }, |
| { |
| "epoch": 17.28, |
| "learning_rate": 0.000392625, |
| "loss": 0.412, |
| "step": 553 |
| }, |
| { |
| "epoch": 17.31, |
| "learning_rate": 0.00039225, |
| "loss": 0.4332, |
| "step": 554 |
| }, |
| { |
| "epoch": 17.34, |
| "learning_rate": 0.0003918749999999999, |
| "loss": 0.3929, |
| "step": 555 |
| }, |
| { |
| "epoch": 17.38, |
| "learning_rate": 0.0003914999999999999, |
| "loss": 0.3933, |
| "step": 556 |
| }, |
| { |
| "epoch": 17.41, |
| "learning_rate": 0.000391125, |
| "loss": 0.4037, |
| "step": 557 |
| }, |
| { |
| "epoch": 17.44, |
| "learning_rate": 0.00039075, |
| "loss": 0.4679, |
| "step": 558 |
| }, |
| { |
| "epoch": 17.47, |
| "learning_rate": 0.000390375, |
| "loss": 0.4333, |
| "step": 559 |
| }, |
| { |
| "epoch": 17.5, |
| "learning_rate": 0.00039, |
| "loss": 0.3785, |
| "step": 560 |
| }, |
| { |
| "epoch": 17.5, |
| "eval_loss": 0.9648825526237488, |
| "eval_mean_accuracy": 0.5533779312497582, |
| "eval_mean_iou": 0.4104480076884441, |
| "eval_overall_accuracy": 0.6964933830395079, |
| "eval_per_category_accuracy": [ |
| 0.707069750381015, |
| 0.7710554403323215, |
| 0.744826490060178, |
| 0.4159998889667664, |
| 0.5508633386479527, |
| 0.5546855436072571, |
| 0.5915950430528899, |
| 0.5151476838910074, |
| 0.4200975129322201, |
| 0.6133331177592535, |
| 0.2742896670778328, |
| 0.4815716982884026 |
| ], |
| "eval_per_category_iou": [ |
| 0.5997489262918438, |
| 0.5787403426302707, |
| 0.5897246173408294, |
| 0.28379961794566494, |
| 0.3905406181776399, |
| 0.38635524838973884, |
| 0.4606979417309078, |
| 0.4181259544507408, |
| 0.243949093813198, |
| 0.4213910285260536, |
| 0.1919286890705549, |
| 0.3603740138938858 |
| ], |
| "eval_runtime": 125.8181, |
| "eval_samples_per_second": 0.906, |
| "eval_steps_per_second": 0.079, |
| "step": 560 |
| }, |
| { |
| "epoch": 17.53, |
| "learning_rate": 0.000389625, |
| "loss": 0.3861, |
| "step": 561 |
| }, |
| { |
| "epoch": 17.56, |
| "learning_rate": 0.00038925, |
| "loss": 0.3942, |
| "step": 562 |
| }, |
| { |
| "epoch": 17.59, |
| "learning_rate": 0.00038887499999999995, |
| "loss": 0.4077, |
| "step": 563 |
| }, |
| { |
| "epoch": 17.62, |
| "learning_rate": 0.00038849999999999996, |
| "loss": 0.4158, |
| "step": 564 |
| }, |
| { |
| "epoch": 17.66, |
| "learning_rate": 0.00038812499999999996, |
| "loss": 0.378, |
| "step": 565 |
| }, |
| { |
| "epoch": 17.69, |
| "learning_rate": 0.00038774999999999997, |
| "loss": 0.452, |
| "step": 566 |
| }, |
| { |
| "epoch": 17.72, |
| "learning_rate": 0.00038737499999999997, |
| "loss": 0.4787, |
| "step": 567 |
| }, |
| { |
| "epoch": 17.75, |
| "learning_rate": 0.000387, |
| "loss": 0.4412, |
| "step": 568 |
| }, |
| { |
| "epoch": 17.78, |
| "learning_rate": 0.000386625, |
| "loss": 0.3697, |
| "step": 569 |
| }, |
| { |
| "epoch": 17.81, |
| "learning_rate": 0.00038625, |
| "loss": 0.3637, |
| "step": 570 |
| }, |
| { |
| "epoch": 17.84, |
| "learning_rate": 0.00038587499999999993, |
| "loss": 0.414, |
| "step": 571 |
| }, |
| { |
| "epoch": 17.88, |
| "learning_rate": 0.00038549999999999994, |
| "loss": 0.3911, |
| "step": 572 |
| }, |
| { |
| "epoch": 17.91, |
| "learning_rate": 0.00038512499999999994, |
| "loss": 0.3696, |
| "step": 573 |
| }, |
| { |
| "epoch": 17.94, |
| "learning_rate": 0.00038474999999999995, |
| "loss": 0.3695, |
| "step": 574 |
| }, |
| { |
| "epoch": 17.97, |
| "learning_rate": 0.00038437499999999995, |
| "loss": 0.4842, |
| "step": 575 |
| }, |
| { |
| "epoch": 18.0, |
| "learning_rate": 0.00038399999999999996, |
| "loss": 0.4701, |
| "step": 576 |
| }, |
| { |
| "epoch": 18.03, |
| "learning_rate": 0.00038362499999999996, |
| "loss": 0.3924, |
| "step": 577 |
| }, |
| { |
| "epoch": 18.06, |
| "learning_rate": 0.00038324999999999996, |
| "loss": 0.3551, |
| "step": 578 |
| }, |
| { |
| "epoch": 18.09, |
| "learning_rate": 0.000382875, |
| "loss": 0.4082, |
| "step": 579 |
| }, |
| { |
| "epoch": 18.12, |
| "learning_rate": 0.0003824999999999999, |
| "loss": 0.3733, |
| "step": 580 |
| }, |
| { |
| "epoch": 18.12, |
| "eval_loss": 1.0153958797454834, |
| "eval_mean_accuracy": 0.5423053726943622, |
| "eval_mean_iou": 0.3962943111689668, |
| "eval_overall_accuracy": 0.6843063622190241, |
| "eval_per_category_accuracy": [ |
| 0.6327200236002443, |
| 0.7950212891251239, |
| 0.722033968917707, |
| 0.3400149894865407, |
| 0.6120719519779906, |
| 0.5697977418804597, |
| 0.6495617563752757, |
| 0.5456487708813033, |
| 0.41224945209976666, |
| 0.504468902830505, |
| 0.19663435492312825, |
| 0.5274412702343 |
| ], |
| "eval_per_category_iou": [ |
| 0.5559925361607938, |
| 0.57438278428341, |
| 0.5822674179792661, |
| 0.25877590254542765, |
| 0.37440670782637575, |
| 0.39002465578168766, |
| 0.44261521238516044, |
| 0.4138787458643415, |
| 0.24997713381827572, |
| 0.39601702066332767, |
| 0.15898387715455906, |
| 0.35820973956497704 |
| ], |
| "eval_runtime": 123.8398, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 580 |
| }, |
| { |
| "epoch": 18.16, |
| "learning_rate": 0.0003821249999999999, |
| "loss": 0.428, |
| "step": 581 |
| }, |
| { |
| "epoch": 18.19, |
| "learning_rate": 0.00038175, |
| "loss": 0.4346, |
| "step": 582 |
| }, |
| { |
| "epoch": 18.22, |
| "learning_rate": 0.000381375, |
| "loss": 0.4228, |
| "step": 583 |
| }, |
| { |
| "epoch": 18.25, |
| "learning_rate": 0.000381, |
| "loss": 0.3848, |
| "step": 584 |
| }, |
| { |
| "epoch": 18.28, |
| "learning_rate": 0.000380625, |
| "loss": 0.4221, |
| "step": 585 |
| }, |
| { |
| "epoch": 18.31, |
| "learning_rate": 0.00038025, |
| "loss": 0.3652, |
| "step": 586 |
| }, |
| { |
| "epoch": 18.34, |
| "learning_rate": 0.000379875, |
| "loss": 0.3645, |
| "step": 587 |
| }, |
| { |
| "epoch": 18.38, |
| "learning_rate": 0.00037949999999999995, |
| "loss": 0.3773, |
| "step": 588 |
| }, |
| { |
| "epoch": 18.41, |
| "learning_rate": 0.00037912499999999996, |
| "loss": 0.3481, |
| "step": 589 |
| }, |
| { |
| "epoch": 18.44, |
| "learning_rate": 0.00037874999999999996, |
| "loss": 0.3516, |
| "step": 590 |
| }, |
| { |
| "epoch": 18.47, |
| "learning_rate": 0.00037837499999999997, |
| "loss": 0.4458, |
| "step": 591 |
| }, |
| { |
| "epoch": 18.5, |
| "learning_rate": 0.00037799999999999997, |
| "loss": 0.3666, |
| "step": 592 |
| }, |
| { |
| "epoch": 18.53, |
| "learning_rate": 0.000377625, |
| "loss": 0.3964, |
| "step": 593 |
| }, |
| { |
| "epoch": 18.56, |
| "learning_rate": 0.00037725, |
| "loss": 0.36, |
| "step": 594 |
| }, |
| { |
| "epoch": 18.59, |
| "learning_rate": 0.000376875, |
| "loss": 0.3631, |
| "step": 595 |
| }, |
| { |
| "epoch": 18.62, |
| "learning_rate": 0.00037649999999999994, |
| "loss": 0.3862, |
| "step": 596 |
| }, |
| { |
| "epoch": 18.66, |
| "learning_rate": 0.00037612499999999994, |
| "loss": 0.3754, |
| "step": 597 |
| }, |
| { |
| "epoch": 18.69, |
| "learning_rate": 0.00037574999999999994, |
| "loss": 0.4267, |
| "step": 598 |
| }, |
| { |
| "epoch": 18.72, |
| "learning_rate": 0.00037537499999999995, |
| "loss": 0.4409, |
| "step": 599 |
| }, |
| { |
| "epoch": 18.75, |
| "learning_rate": 0.00037499999999999995, |
| "loss": 0.3695, |
| "step": 600 |
| }, |
| { |
| "epoch": 18.75, |
| "eval_loss": 1.0076940059661865, |
| "eval_mean_accuracy": 0.5395021806894327, |
| "eval_mean_iou": 0.406199651624731, |
| "eval_overall_accuracy": 0.6966659478973924, |
| "eval_per_category_accuracy": [ |
| 0.7092971776181227, |
| 0.7965909912228947, |
| 0.683093642092215, |
| 0.3118715345486846, |
| 0.5729537871221437, |
| 0.5305219950714781, |
| 0.561815416540056, |
| 0.6014393915854415, |
| 0.37046928653185807, |
| 0.565621706084561, |
| 0.3054240265921466, |
| 0.4649272132635901 |
| ], |
| "eval_per_category_iou": [ |
| 0.5993953661114904, |
| 0.582966874261397, |
| 0.5743266262263335, |
| 0.26181307638083806, |
| 0.3892571447447381, |
| 0.38557275607124863, |
| 0.44736254850215323, |
| 0.4196338996431097, |
| 0.24949395207371344, |
| 0.42156653903714214, |
| 0.2034758883947415, |
| 0.33953114804986584 |
| ], |
| "eval_runtime": 124.2803, |
| "eval_samples_per_second": 0.917, |
| "eval_steps_per_second": 0.08, |
| "step": 600 |
| }, |
| { |
| "epoch": 18.78, |
| "learning_rate": 0.00037462499999999996, |
| "loss": 0.5011, |
| "step": 601 |
| }, |
| { |
| "epoch": 18.81, |
| "learning_rate": 0.00037424999999999996, |
| "loss": 0.3674, |
| "step": 602 |
| }, |
| { |
| "epoch": 18.84, |
| "learning_rate": 0.00037387499999999997, |
| "loss": 0.4018, |
| "step": 603 |
| }, |
| { |
| "epoch": 18.88, |
| "learning_rate": 0.0003735, |
| "loss": 0.3229, |
| "step": 604 |
| }, |
| { |
| "epoch": 18.91, |
| "learning_rate": 0.0003731249999999999, |
| "loss": 0.4203, |
| "step": 605 |
| }, |
| { |
| "epoch": 18.94, |
| "learning_rate": 0.0003727499999999999, |
| "loss": 0.2997, |
| "step": 606 |
| }, |
| { |
| "epoch": 18.97, |
| "learning_rate": 0.00037237499999999993, |
| "loss": 0.3523, |
| "step": 607 |
| }, |
| { |
| "epoch": 19.0, |
| "learning_rate": 0.000372, |
| "loss": 0.3813, |
| "step": 608 |
| }, |
| { |
| "epoch": 19.03, |
| "learning_rate": 0.000371625, |
| "loss": 0.3486, |
| "step": 609 |
| }, |
| { |
| "epoch": 19.06, |
| "learning_rate": 0.00037125, |
| "loss": 0.3432, |
| "step": 610 |
| }, |
| { |
| "epoch": 19.09, |
| "learning_rate": 0.000370875, |
| "loss": 0.3893, |
| "step": 611 |
| }, |
| { |
| "epoch": 19.12, |
| "learning_rate": 0.0003705, |
| "loss": 0.3956, |
| "step": 612 |
| }, |
| { |
| "epoch": 19.16, |
| "learning_rate": 0.00037012499999999996, |
| "loss": 0.297, |
| "step": 613 |
| }, |
| { |
| "epoch": 19.19, |
| "learning_rate": 0.00036974999999999996, |
| "loss": 0.3976, |
| "step": 614 |
| }, |
| { |
| "epoch": 19.22, |
| "learning_rate": 0.00036937499999999997, |
| "loss": 0.3581, |
| "step": 615 |
| }, |
| { |
| "epoch": 19.25, |
| "learning_rate": 0.00036899999999999997, |
| "loss": 0.3371, |
| "step": 616 |
| }, |
| { |
| "epoch": 19.28, |
| "learning_rate": 0.000368625, |
| "loss": 0.347, |
| "step": 617 |
| }, |
| { |
| "epoch": 19.31, |
| "learning_rate": 0.00036825, |
| "loss": 0.3433, |
| "step": 618 |
| }, |
| { |
| "epoch": 19.34, |
| "learning_rate": 0.000367875, |
| "loss": 0.394, |
| "step": 619 |
| }, |
| { |
| "epoch": 19.38, |
| "learning_rate": 0.0003675, |
| "loss": 0.3537, |
| "step": 620 |
| }, |
| { |
| "epoch": 19.38, |
| "eval_loss": 1.0441322326660156, |
| "eval_mean_accuracy": 0.5415450698238194, |
| "eval_mean_iou": 0.4053702230412377, |
| "eval_overall_accuracy": 0.6971735703317743, |
| "eval_per_category_accuracy": [ |
| 0.6794642779789252, |
| 0.7984741324962451, |
| 0.7563555821630794, |
| 0.33012262232739537, |
| 0.5992540624423883, |
| 0.6299598241991152, |
| 0.5696072889983296, |
| 0.42462266061366466, |
| 0.3673837431124282, |
| 0.5740040607898044, |
| 0.29494694119915643, |
| 0.47434564156530173 |
| ], |
| "eval_per_category_iou": [ |
| 0.5883688281257252, |
| 0.5826365522664997, |
| 0.6029757945416013, |
| 0.26803659014139664, |
| 0.3681170025561952, |
| 0.39479223645889366, |
| 0.45552532308591115, |
| 0.3849501261295565, |
| 0.24765423717662116, |
| 0.4254810071381738, |
| 0.20019112713070394, |
| 0.345713851743574 |
| ], |
| "eval_runtime": 124.331, |
| "eval_samples_per_second": 0.917, |
| "eval_steps_per_second": 0.08, |
| "step": 620 |
| }, |
| { |
| "epoch": 19.41, |
| "learning_rate": 0.00036712499999999994, |
| "loss": 0.4249, |
| "step": 621 |
| }, |
| { |
| "epoch": 19.44, |
| "learning_rate": 0.00036674999999999994, |
| "loss": 0.4018, |
| "step": 622 |
| }, |
| { |
| "epoch": 19.47, |
| "learning_rate": 0.00036637499999999995, |
| "loss": 0.4046, |
| "step": 623 |
| }, |
| { |
| "epoch": 19.5, |
| "learning_rate": 0.00036599999999999995, |
| "loss": 0.4108, |
| "step": 624 |
| }, |
| { |
| "epoch": 19.53, |
| "learning_rate": 0.00036562499999999996, |
| "loss": 0.3917, |
| "step": 625 |
| }, |
| { |
| "epoch": 19.56, |
| "learning_rate": 0.00036524999999999996, |
| "loss": 0.4627, |
| "step": 626 |
| }, |
| { |
| "epoch": 19.59, |
| "learning_rate": 0.00036487499999999996, |
| "loss": 0.3753, |
| "step": 627 |
| }, |
| { |
| "epoch": 19.62, |
| "learning_rate": 0.00036449999999999997, |
| "loss": 0.3768, |
| "step": 628 |
| }, |
| { |
| "epoch": 19.66, |
| "learning_rate": 0.000364125, |
| "loss": 0.4406, |
| "step": 629 |
| }, |
| { |
| "epoch": 19.69, |
| "learning_rate": 0.0003637499999999999, |
| "loss": 0.3537, |
| "step": 630 |
| }, |
| { |
| "epoch": 19.72, |
| "learning_rate": 0.00036337499999999993, |
| "loss": 0.347, |
| "step": 631 |
| }, |
| { |
| "epoch": 19.75, |
| "learning_rate": 0.00036299999999999993, |
| "loss": 0.3625, |
| "step": 632 |
| }, |
| { |
| "epoch": 19.78, |
| "learning_rate": 0.000362625, |
| "loss": 0.4063, |
| "step": 633 |
| }, |
| { |
| "epoch": 19.81, |
| "learning_rate": 0.00036225, |
| "loss": 0.3952, |
| "step": 634 |
| }, |
| { |
| "epoch": 19.84, |
| "learning_rate": 0.000361875, |
| "loss": 0.3706, |
| "step": 635 |
| }, |
| { |
| "epoch": 19.88, |
| "learning_rate": 0.0003615, |
| "loss": 0.4143, |
| "step": 636 |
| }, |
| { |
| "epoch": 19.91, |
| "learning_rate": 0.000361125, |
| "loss": 0.3519, |
| "step": 637 |
| }, |
| { |
| "epoch": 19.94, |
| "learning_rate": 0.00036074999999999996, |
| "loss": 0.4351, |
| "step": 638 |
| }, |
| { |
| "epoch": 19.97, |
| "learning_rate": 0.00036037499999999996, |
| "loss": 0.3982, |
| "step": 639 |
| }, |
| { |
| "epoch": 20.0, |
| "learning_rate": 0.00035999999999999997, |
| "loss": 0.3443, |
| "step": 640 |
| }, |
| { |
| "epoch": 20.0, |
| "eval_loss": 1.1258885860443115, |
| "eval_mean_accuracy": 0.5394273904447178, |
| "eval_mean_iou": 0.3833654483635454, |
| "eval_overall_accuracy": 0.6722515507748252, |
| "eval_per_category_accuracy": [ |
| 0.5770503483135855, |
| 0.7866469575380061, |
| 0.7340458772431193, |
| 0.36867891270705966, |
| 0.6653187657387355, |
| 0.624125804332599, |
| 0.6725752539955192, |
| 0.5482641906074182, |
| 0.34888822835239525, |
| 0.5821668428637069, |
| 0.16921885569291703, |
| 0.3961486479515512 |
| ], |
| "eval_per_category_iou": [ |
| 0.5224325457567354, |
| 0.5638760112021677, |
| 0.5784232887631695, |
| 0.2501383065626758, |
| 0.3528750625901244, |
| 0.38667239278149773, |
| 0.43968983934101924, |
| 0.40781501795593983, |
| 0.22104279390063944, |
| 0.40770197560634414, |
| 0.14258707622474112, |
| 0.3271310696774908 |
| ], |
| "eval_runtime": 123.7678, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 640 |
| }, |
| { |
| "epoch": 20.03, |
| "learning_rate": 0.000359625, |
| "loss": 0.343, |
| "step": 641 |
| }, |
| { |
| "epoch": 20.06, |
| "learning_rate": 0.00035925, |
| "loss": 0.3623, |
| "step": 642 |
| }, |
| { |
| "epoch": 20.09, |
| "learning_rate": 0.000358875, |
| "loss": 0.4264, |
| "step": 643 |
| }, |
| { |
| "epoch": 20.12, |
| "learning_rate": 0.0003585, |
| "loss": 0.3634, |
| "step": 644 |
| }, |
| { |
| "epoch": 20.16, |
| "learning_rate": 0.000358125, |
| "loss": 0.3663, |
| "step": 645 |
| }, |
| { |
| "epoch": 20.19, |
| "learning_rate": 0.00035774999999999994, |
| "loss": 0.3666, |
| "step": 646 |
| }, |
| { |
| "epoch": 20.22, |
| "learning_rate": 0.00035737499999999995, |
| "loss": 0.3391, |
| "step": 647 |
| }, |
| { |
| "epoch": 20.25, |
| "learning_rate": 0.00035699999999999995, |
| "loss": 0.3908, |
| "step": 648 |
| }, |
| { |
| "epoch": 20.28, |
| "learning_rate": 0.00035662499999999995, |
| "loss": 0.3773, |
| "step": 649 |
| }, |
| { |
| "epoch": 20.31, |
| "learning_rate": 0.00035624999999999996, |
| "loss": 0.4099, |
| "step": 650 |
| }, |
| { |
| "epoch": 20.34, |
| "learning_rate": 0.00035587499999999996, |
| "loss": 0.3897, |
| "step": 651 |
| }, |
| { |
| "epoch": 20.38, |
| "learning_rate": 0.00035549999999999997, |
| "loss": 0.3473, |
| "step": 652 |
| }, |
| { |
| "epoch": 20.41, |
| "learning_rate": 0.00035512499999999997, |
| "loss": 0.3624, |
| "step": 653 |
| }, |
| { |
| "epoch": 20.44, |
| "learning_rate": 0.00035475, |
| "loss": 0.3227, |
| "step": 654 |
| }, |
| { |
| "epoch": 20.47, |
| "learning_rate": 0.0003543749999999999, |
| "loss": 0.3288, |
| "step": 655 |
| }, |
| { |
| "epoch": 20.5, |
| "learning_rate": 0.00035399999999999993, |
| "loss": 0.3624, |
| "step": 656 |
| }, |
| { |
| "epoch": 20.53, |
| "learning_rate": 0.00035362499999999994, |
| "loss": 0.3919, |
| "step": 657 |
| }, |
| { |
| "epoch": 20.56, |
| "learning_rate": 0.00035324999999999994, |
| "loss": 0.2955, |
| "step": 658 |
| }, |
| { |
| "epoch": 20.59, |
| "learning_rate": 0.000352875, |
| "loss": 0.4382, |
| "step": 659 |
| }, |
| { |
| "epoch": 20.62, |
| "learning_rate": 0.0003525, |
| "loss": 0.3643, |
| "step": 660 |
| }, |
| { |
| "epoch": 20.62, |
| "eval_loss": 0.9908437132835388, |
| "eval_mean_accuracy": 0.5424145852163237, |
| "eval_mean_iou": 0.4064449473686127, |
| "eval_overall_accuracy": 0.6996373293692606, |
| "eval_per_category_accuracy": [ |
| 0.7225005099044816, |
| 0.7738001553209554, |
| 0.7652851141081354, |
| 0.3469476270116099, |
| 0.6430865040406639, |
| 0.5881944116803792, |
| 0.6214266258783372, |
| 0.53267131749224, |
| 0.23419518557624908, |
| 0.5568341068791949, |
| 0.26203732327761675, |
| 0.46199614142602163 |
| ], |
| "eval_per_category_iou": [ |
| 0.6025222099167149, |
| 0.5792443438352654, |
| 0.6025714767174072, |
| 0.2566168186402704, |
| 0.3824601734512165, |
| 0.39593947692762466, |
| 0.45743661836347865, |
| 0.43561784993270525, |
| 0.19584375979654717, |
| 0.4124189378991299, |
| 0.20226287960762263, |
| 0.35440482333537 |
| ], |
| "eval_runtime": 123.9944, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 660 |
| }, |
| { |
| "epoch": 20.66, |
| "learning_rate": 0.000352125, |
| "loss": 0.3625, |
| "step": 661 |
| }, |
| { |
| "epoch": 20.69, |
| "learning_rate": 0.00035175, |
| "loss": 0.3323, |
| "step": 662 |
| }, |
| { |
| "epoch": 20.72, |
| "learning_rate": 0.00035137499999999996, |
| "loss": 0.3515, |
| "step": 663 |
| }, |
| { |
| "epoch": 20.75, |
| "learning_rate": 0.00035099999999999997, |
| "loss": 0.456, |
| "step": 664 |
| }, |
| { |
| "epoch": 20.78, |
| "learning_rate": 0.00035062499999999997, |
| "loss": 0.3541, |
| "step": 665 |
| }, |
| { |
| "epoch": 20.81, |
| "learning_rate": 0.00035025, |
| "loss": 0.3759, |
| "step": 666 |
| }, |
| { |
| "epoch": 20.84, |
| "learning_rate": 0.000349875, |
| "loss": 0.3632, |
| "step": 667 |
| }, |
| { |
| "epoch": 20.88, |
| "learning_rate": 0.0003495, |
| "loss": 0.3571, |
| "step": 668 |
| }, |
| { |
| "epoch": 20.91, |
| "learning_rate": 0.000349125, |
| "loss": 0.3406, |
| "step": 669 |
| }, |
| { |
| "epoch": 20.94, |
| "learning_rate": 0.00034875, |
| "loss": 0.3889, |
| "step": 670 |
| }, |
| { |
| "epoch": 20.97, |
| "learning_rate": 0.00034837499999999994, |
| "loss": 0.3544, |
| "step": 671 |
| }, |
| { |
| "epoch": 21.0, |
| "learning_rate": 0.00034799999999999995, |
| "loss": 0.3768, |
| "step": 672 |
| }, |
| { |
| "epoch": 21.03, |
| "learning_rate": 0.00034762499999999995, |
| "loss": 0.4055, |
| "step": 673 |
| }, |
| { |
| "epoch": 21.06, |
| "learning_rate": 0.00034724999999999996, |
| "loss": 0.3344, |
| "step": 674 |
| }, |
| { |
| "epoch": 21.09, |
| "learning_rate": 0.00034687499999999996, |
| "loss": 0.3785, |
| "step": 675 |
| }, |
| { |
| "epoch": 21.12, |
| "learning_rate": 0.00034649999999999997, |
| "loss": 0.3469, |
| "step": 676 |
| }, |
| { |
| "epoch": 21.16, |
| "learning_rate": 0.00034612499999999997, |
| "loss": 0.3359, |
| "step": 677 |
| }, |
| { |
| "epoch": 21.19, |
| "learning_rate": 0.00034575, |
| "loss": 0.4219, |
| "step": 678 |
| }, |
| { |
| "epoch": 21.22, |
| "learning_rate": 0.000345375, |
| "loss": 0.3693, |
| "step": 679 |
| }, |
| { |
| "epoch": 21.25, |
| "learning_rate": 0.00034499999999999993, |
| "loss": 0.3621, |
| "step": 680 |
| }, |
| { |
| "epoch": 21.25, |
| "eval_loss": 1.043190360069275, |
| "eval_mean_accuracy": 0.547674649061508, |
| "eval_mean_iou": 0.40279079404193263, |
| "eval_overall_accuracy": 0.6930079878422252, |
| "eval_per_category_accuracy": [ |
| 0.7290935124437985, |
| 0.764183466289645, |
| 0.6980677934471109, |
| 0.38046231462654667, |
| 0.606832966307204, |
| 0.5785641959940335, |
| 0.6419674753483041, |
| 0.5622380964115927, |
| 0.40316584296780034, |
| 0.5970305018199494, |
| 0.21820986615163768, |
| 0.39227975693047346 |
| ], |
| "eval_per_category_iou": [ |
| 0.5921803600550819, |
| 0.5737549467190497, |
| 0.5882262700972748, |
| 0.25031674827928635, |
| 0.387075222879975, |
| 0.3861233258773136, |
| 0.4502504428023092, |
| 0.43198146097708556, |
| 0.2470879963457209, |
| 0.43134736160667353, |
| 0.16907260503005941, |
| 0.3260727878333622 |
| ], |
| "eval_runtime": 123.9243, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 680 |
| }, |
| { |
| "epoch": 21.28, |
| "learning_rate": 0.00034462499999999993, |
| "loss": 0.3123, |
| "step": 681 |
| }, |
| { |
| "epoch": 21.31, |
| "learning_rate": 0.00034424999999999994, |
| "loss": 0.3967, |
| "step": 682 |
| }, |
| { |
| "epoch": 21.34, |
| "learning_rate": 0.00034387499999999994, |
| "loss": 0.3347, |
| "step": 683 |
| }, |
| { |
| "epoch": 21.38, |
| "learning_rate": 0.0003435, |
| "loss": 0.3405, |
| "step": 684 |
| }, |
| { |
| "epoch": 21.41, |
| "learning_rate": 0.000343125, |
| "loss": 0.3754, |
| "step": 685 |
| }, |
| { |
| "epoch": 21.44, |
| "learning_rate": 0.00034275, |
| "loss": 0.3175, |
| "step": 686 |
| }, |
| { |
| "epoch": 21.47, |
| "learning_rate": 0.000342375, |
| "loss": 0.3332, |
| "step": 687 |
| }, |
| { |
| "epoch": 21.5, |
| "learning_rate": 0.00034199999999999996, |
| "loss": 0.3551, |
| "step": 688 |
| }, |
| { |
| "epoch": 21.53, |
| "learning_rate": 0.00034162499999999997, |
| "loss": 0.309, |
| "step": 689 |
| }, |
| { |
| "epoch": 21.56, |
| "learning_rate": 0.00034125, |
| "loss": 0.3129, |
| "step": 690 |
| }, |
| { |
| "epoch": 21.59, |
| "learning_rate": 0.000340875, |
| "loss": 0.3966, |
| "step": 691 |
| }, |
| { |
| "epoch": 21.62, |
| "learning_rate": 0.0003405, |
| "loss": 0.3929, |
| "step": 692 |
| }, |
| { |
| "epoch": 21.66, |
| "learning_rate": 0.000340125, |
| "loss": 0.3579, |
| "step": 693 |
| }, |
| { |
| "epoch": 21.69, |
| "learning_rate": 0.00033975, |
| "loss": 0.3499, |
| "step": 694 |
| }, |
| { |
| "epoch": 21.72, |
| "learning_rate": 0.000339375, |
| "loss": 0.4075, |
| "step": 695 |
| }, |
| { |
| "epoch": 21.75, |
| "learning_rate": 0.00033899999999999995, |
| "loss": 0.4487, |
| "step": 696 |
| }, |
| { |
| "epoch": 21.78, |
| "learning_rate": 0.00033862499999999995, |
| "loss": 0.4009, |
| "step": 697 |
| }, |
| { |
| "epoch": 21.81, |
| "learning_rate": 0.00033824999999999995, |
| "loss": 0.3475, |
| "step": 698 |
| }, |
| { |
| "epoch": 21.84, |
| "learning_rate": 0.00033787499999999996, |
| "loss": 0.3563, |
| "step": 699 |
| }, |
| { |
| "epoch": 21.88, |
| "learning_rate": 0.00033749999999999996, |
| "loss": 0.3579, |
| "step": 700 |
| }, |
| { |
| "epoch": 21.88, |
| "eval_loss": 1.0316771268844604, |
| "eval_mean_accuracy": 0.5426885321379415, |
| "eval_mean_iou": 0.4060680967213626, |
| "eval_overall_accuracy": 0.6925102702358312, |
| "eval_per_category_accuracy": [ |
| 0.6682704612664221, |
| 0.7819589519990984, |
| 0.7712678120520859, |
| 0.43665554021137953, |
| 0.5480534109119198, |
| 0.4661935334018052, |
| 0.6375354444390214, |
| 0.4838558636938812, |
| 0.41139765578556026, |
| 0.5998280970563186, |
| 0.25927342958498745, |
| 0.4479721852528191 |
| ], |
| "eval_per_category_iou": [ |
| 0.5745877754266178, |
| 0.5764814441641772, |
| 0.5915648537689412, |
| 0.2725303996621658, |
| 0.38916557335542973, |
| 0.37087793505171757, |
| 0.4601192312521554, |
| 0.4147331530843154, |
| 0.2611279599232933, |
| 0.42749856798813207, |
| 0.18955678781938534, |
| 0.3445734791600205 |
| ], |
| "eval_runtime": 124.0103, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 700 |
| }, |
| { |
| "epoch": 21.91, |
| "learning_rate": 0.00033712499999999997, |
| "loss": 0.3044, |
| "step": 701 |
| }, |
| { |
| "epoch": 21.94, |
| "learning_rate": 0.00033674999999999997, |
| "loss": 0.3342, |
| "step": 702 |
| }, |
| { |
| "epoch": 21.97, |
| "learning_rate": 0.000336375, |
| "loss": 0.3963, |
| "step": 703 |
| }, |
| { |
| "epoch": 22.0, |
| "learning_rate": 0.000336, |
| "loss": 0.3001, |
| "step": 704 |
| }, |
| { |
| "epoch": 22.03, |
| "learning_rate": 0.00033562499999999993, |
| "loss": 0.3634, |
| "step": 705 |
| }, |
| { |
| "epoch": 22.06, |
| "learning_rate": 0.00033524999999999994, |
| "loss": 0.3505, |
| "step": 706 |
| }, |
| { |
| "epoch": 22.09, |
| "learning_rate": 0.00033487499999999994, |
| "loss": 0.3528, |
| "step": 707 |
| }, |
| { |
| "epoch": 22.12, |
| "learning_rate": 0.00033449999999999994, |
| "loss": 0.3699, |
| "step": 708 |
| }, |
| { |
| "epoch": 22.16, |
| "learning_rate": 0.00033412499999999995, |
| "loss": 0.3382, |
| "step": 709 |
| }, |
| { |
| "epoch": 22.19, |
| "learning_rate": 0.00033375, |
| "loss": 0.2887, |
| "step": 710 |
| }, |
| { |
| "epoch": 22.22, |
| "learning_rate": 0.000333375, |
| "loss": 0.3868, |
| "step": 711 |
| }, |
| { |
| "epoch": 22.25, |
| "learning_rate": 0.000333, |
| "loss": 0.3529, |
| "step": 712 |
| }, |
| { |
| "epoch": 22.28, |
| "learning_rate": 0.00033262499999999997, |
| "loss": 0.3106, |
| "step": 713 |
| }, |
| { |
| "epoch": 22.31, |
| "learning_rate": 0.00033224999999999997, |
| "loss": 0.3625, |
| "step": 714 |
| }, |
| { |
| "epoch": 22.34, |
| "learning_rate": 0.000331875, |
| "loss": 0.3509, |
| "step": 715 |
| }, |
| { |
| "epoch": 22.38, |
| "learning_rate": 0.0003315, |
| "loss": 0.3982, |
| "step": 716 |
| }, |
| { |
| "epoch": 22.41, |
| "learning_rate": 0.000331125, |
| "loss": 0.3949, |
| "step": 717 |
| }, |
| { |
| "epoch": 22.44, |
| "learning_rate": 0.00033075, |
| "loss": 0.3126, |
| "step": 718 |
| }, |
| { |
| "epoch": 22.47, |
| "learning_rate": 0.000330375, |
| "loss": 0.3818, |
| "step": 719 |
| }, |
| { |
| "epoch": 22.5, |
| "learning_rate": 0.00033, |
| "loss": 0.3442, |
| "step": 720 |
| }, |
| { |
| "epoch": 22.5, |
| "eval_loss": 1.0697511434555054, |
| "eval_mean_accuracy": 0.5536363964270905, |
| "eval_mean_iou": 0.40210358786153266, |
| "eval_overall_accuracy": 0.6880487475478858, |
| "eval_per_category_accuracy": [ |
| 0.6469024899638023, |
| 0.7827470368800764, |
| 0.7353876041215156, |
| 0.42743631203114485, |
| 0.6404467521732876, |
| 0.5755011087359684, |
| 0.725127489770117, |
| 0.45122151137187927, |
| 0.3425862665590091, |
| 0.584515904933917, |
| 0.28915592584317124, |
| 0.442608354741197 |
| ], |
| "eval_per_category_iou": [ |
| 0.5693479729109744, |
| 0.5747118078769436, |
| 0.592137509150473, |
| 0.2712197269925143, |
| 0.3839531768269503, |
| 0.39685465413833965, |
| 0.4424154544358594, |
| 0.400306458343433, |
| 0.23499356377944608, |
| 0.42245057429150734, |
| 0.20232368877561946, |
| 0.33452846681633125 |
| ], |
| "eval_runtime": 123.3393, |
| "eval_samples_per_second": 0.924, |
| "eval_steps_per_second": 0.081, |
| "step": 720 |
| }, |
| { |
| "epoch": 22.53, |
| "learning_rate": 0.00032962499999999995, |
| "loss": 0.3743, |
| "step": 721 |
| }, |
| { |
| "epoch": 22.56, |
| "learning_rate": 0.00032924999999999995, |
| "loss": 0.3076, |
| "step": 722 |
| }, |
| { |
| "epoch": 22.59, |
| "learning_rate": 0.00032887499999999996, |
| "loss": 0.3052, |
| "step": 723 |
| }, |
| { |
| "epoch": 22.62, |
| "learning_rate": 0.00032849999999999996, |
| "loss": 0.3788, |
| "step": 724 |
| }, |
| { |
| "epoch": 22.66, |
| "learning_rate": 0.00032812499999999997, |
| "loss": 0.3889, |
| "step": 725 |
| }, |
| { |
| "epoch": 22.69, |
| "learning_rate": 0.00032774999999999997, |
| "loss": 0.305, |
| "step": 726 |
| }, |
| { |
| "epoch": 22.72, |
| "learning_rate": 0.000327375, |
| "loss": 0.3152, |
| "step": 727 |
| }, |
| { |
| "epoch": 22.75, |
| "learning_rate": 0.000327, |
| "loss": 0.3355, |
| "step": 728 |
| }, |
| { |
| "epoch": 22.78, |
| "learning_rate": 0.000326625, |
| "loss": 0.367, |
| "step": 729 |
| }, |
| { |
| "epoch": 22.81, |
| "learning_rate": 0.00032624999999999993, |
| "loss": 0.3734, |
| "step": 730 |
| }, |
| { |
| "epoch": 22.84, |
| "learning_rate": 0.00032587499999999994, |
| "loss": 0.3064, |
| "step": 731 |
| }, |
| { |
| "epoch": 22.88, |
| "learning_rate": 0.00032549999999999994, |
| "loss": 0.3818, |
| "step": 732 |
| }, |
| { |
| "epoch": 22.91, |
| "learning_rate": 0.00032512499999999995, |
| "loss": 0.3606, |
| "step": 733 |
| }, |
| { |
| "epoch": 22.94, |
| "learning_rate": 0.00032474999999999995, |
| "loss": 0.3496, |
| "step": 734 |
| }, |
| { |
| "epoch": 22.97, |
| "learning_rate": 0.000324375, |
| "loss": 0.3293, |
| "step": 735 |
| }, |
| { |
| "epoch": 23.0, |
| "learning_rate": 0.000324, |
| "loss": 0.3207, |
| "step": 736 |
| }, |
| { |
| "epoch": 23.03, |
| "learning_rate": 0.000323625, |
| "loss": 0.3937, |
| "step": 737 |
| }, |
| { |
| "epoch": 23.06, |
| "learning_rate": 0.0003232499999999999, |
| "loss": 0.3863, |
| "step": 738 |
| }, |
| { |
| "epoch": 23.09, |
| "learning_rate": 0.000322875, |
| "loss": 0.3356, |
| "step": 739 |
| }, |
| { |
| "epoch": 23.12, |
| "learning_rate": 0.0003225, |
| "loss": 0.3518, |
| "step": 740 |
| }, |
| { |
| "epoch": 23.12, |
| "eval_loss": 1.0315091609954834, |
| "eval_mean_accuracy": 0.547964758210377, |
| "eval_mean_iou": 0.4089135486955994, |
| "eval_overall_accuracy": 0.6973423204923931, |
| "eval_per_category_accuracy": [ |
| 0.7417658686019102, |
| 0.7614223396031409, |
| 0.7227599053657375, |
| 0.33566387464347924, |
| 0.5676904906129775, |
| 0.5667382838727004, |
| 0.6773749151773836, |
| 0.48031363607629185, |
| 0.39033122454593044, |
| 0.5798834961178412, |
| 0.2758873996443362, |
| 0.4757456642627957 |
| ], |
| "eval_per_category_iou": [ |
| 0.6109214860216081, |
| 0.5753985745846153, |
| 0.5959716055201187, |
| 0.2711044101862776, |
| 0.3857312369399852, |
| 0.40771815585134435, |
| 0.44582471313053124, |
| 0.41556190248410874, |
| 0.24282560097149009, |
| 0.4205023278330163, |
| 0.19415484827075183, |
| 0.3412477225533452 |
| ], |
| "eval_runtime": 124.8036, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 740 |
| }, |
| { |
| "epoch": 23.16, |
| "learning_rate": 0.000322125, |
| "loss": 0.3215, |
| "step": 741 |
| }, |
| { |
| "epoch": 23.19, |
| "learning_rate": 0.00032175, |
| "loss": 0.2858, |
| "step": 742 |
| }, |
| { |
| "epoch": 23.22, |
| "learning_rate": 0.000321375, |
| "loss": 0.372, |
| "step": 743 |
| }, |
| { |
| "epoch": 23.25, |
| "learning_rate": 0.000321, |
| "loss": 0.3286, |
| "step": 744 |
| }, |
| { |
| "epoch": 23.28, |
| "learning_rate": 0.000320625, |
| "loss": 0.2929, |
| "step": 745 |
| }, |
| { |
| "epoch": 23.31, |
| "learning_rate": 0.00032024999999999995, |
| "loss": 0.2972, |
| "step": 746 |
| }, |
| { |
| "epoch": 23.34, |
| "learning_rate": 0.00031987499999999996, |
| "loss": 0.3627, |
| "step": 747 |
| }, |
| { |
| "epoch": 23.38, |
| "learning_rate": 0.00031949999999999996, |
| "loss": 0.3223, |
| "step": 748 |
| }, |
| { |
| "epoch": 23.41, |
| "learning_rate": 0.00031912499999999996, |
| "loss": 0.3214, |
| "step": 749 |
| }, |
| { |
| "epoch": 23.44, |
| "learning_rate": 0.00031874999999999997, |
| "loss": 0.2965, |
| "step": 750 |
| }, |
| { |
| "epoch": 23.47, |
| "learning_rate": 0.000318375, |
| "loss": 0.3354, |
| "step": 751 |
| }, |
| { |
| "epoch": 23.5, |
| "learning_rate": 0.000318, |
| "loss": 0.3251, |
| "step": 752 |
| }, |
| { |
| "epoch": 23.53, |
| "learning_rate": 0.000317625, |
| "loss": 0.361, |
| "step": 753 |
| }, |
| { |
| "epoch": 23.56, |
| "learning_rate": 0.00031725, |
| "loss": 0.3347, |
| "step": 754 |
| }, |
| { |
| "epoch": 23.59, |
| "learning_rate": 0.00031687499999999994, |
| "loss": 0.3174, |
| "step": 755 |
| }, |
| { |
| "epoch": 23.62, |
| "learning_rate": 0.00031649999999999994, |
| "loss": 0.3501, |
| "step": 756 |
| }, |
| { |
| "epoch": 23.66, |
| "learning_rate": 0.00031612499999999995, |
| "loss": 0.3335, |
| "step": 757 |
| }, |
| { |
| "epoch": 23.69, |
| "learning_rate": 0.00031574999999999995, |
| "loss": 0.3361, |
| "step": 758 |
| }, |
| { |
| "epoch": 23.72, |
| "learning_rate": 0.00031537499999999995, |
| "loss": 0.3746, |
| "step": 759 |
| }, |
| { |
| "epoch": 23.75, |
| "learning_rate": 0.00031499999999999996, |
| "loss": 0.3478, |
| "step": 760 |
| }, |
| { |
| "epoch": 23.75, |
| "eval_loss": 1.0356502532958984, |
| "eval_mean_accuracy": 0.5445703040180192, |
| "eval_mean_iou": 0.4096686929361411, |
| "eval_overall_accuracy": 0.7005514512982285, |
| "eval_per_category_accuracy": [ |
| 0.7313842280413251, |
| 0.7757354421367998, |
| 0.7531878854218892, |
| 0.33439740182233296, |
| 0.587050021575869, |
| 0.6033810095848501, |
| 0.6232757982772547, |
| 0.5013237816762044, |
| 0.37167599797698375, |
| 0.560122480847373, |
| 0.26734779454750807, |
| 0.4259618063078399 |
| ], |
| "eval_per_category_iou": [ |
| 0.6070527975647682, |
| 0.5826104929243641, |
| 0.5947798650585946, |
| 0.2745643249279929, |
| 0.37872702487580395, |
| 0.4054199048480666, |
| 0.4566098758568507, |
| 0.4313106787166809, |
| 0.2448435643737972, |
| 0.41498880567313184, |
| 0.19164364716646226, |
| 0.33347333324717954 |
| ], |
| "eval_runtime": 124.5129, |
| "eval_samples_per_second": 0.916, |
| "eval_steps_per_second": 0.08, |
| "step": 760 |
| }, |
| { |
| "epoch": 23.78, |
| "learning_rate": 0.000314625, |
| "loss": 0.3457, |
| "step": 761 |
| }, |
| { |
| "epoch": 23.81, |
| "learning_rate": 0.00031425, |
| "loss": 0.3957, |
| "step": 762 |
| }, |
| { |
| "epoch": 23.84, |
| "learning_rate": 0.0003138749999999999, |
| "loss": 0.4193, |
| "step": 763 |
| }, |
| { |
| "epoch": 23.88, |
| "learning_rate": 0.0003135, |
| "loss": 0.3332, |
| "step": 764 |
| }, |
| { |
| "epoch": 23.91, |
| "learning_rate": 0.000313125, |
| "loss": 0.3274, |
| "step": 765 |
| }, |
| { |
| "epoch": 23.94, |
| "learning_rate": 0.00031275, |
| "loss": 0.3406, |
| "step": 766 |
| }, |
| { |
| "epoch": 23.97, |
| "learning_rate": 0.000312375, |
| "loss": 0.362, |
| "step": 767 |
| }, |
| { |
| "epoch": 24.0, |
| "learning_rate": 0.000312, |
| "loss": 0.2929, |
| "step": 768 |
| }, |
| { |
| "epoch": 24.03, |
| "learning_rate": 0.000311625, |
| "loss": 0.3477, |
| "step": 769 |
| }, |
| { |
| "epoch": 24.06, |
| "learning_rate": 0.00031125, |
| "loss": 0.3376, |
| "step": 770 |
| }, |
| { |
| "epoch": 24.09, |
| "learning_rate": 0.00031087499999999995, |
| "loss": 0.3083, |
| "step": 771 |
| }, |
| { |
| "epoch": 24.12, |
| "learning_rate": 0.00031049999999999996, |
| "loss": 0.2948, |
| "step": 772 |
| }, |
| { |
| "epoch": 24.16, |
| "learning_rate": 0.00031012499999999996, |
| "loss": 0.3991, |
| "step": 773 |
| }, |
| { |
| "epoch": 24.19, |
| "learning_rate": 0.00030974999999999997, |
| "loss": 0.2986, |
| "step": 774 |
| }, |
| { |
| "epoch": 24.22, |
| "learning_rate": 0.00030937499999999997, |
| "loss": 0.3202, |
| "step": 775 |
| }, |
| { |
| "epoch": 24.25, |
| "learning_rate": 0.000309, |
| "loss": 0.2929, |
| "step": 776 |
| }, |
| { |
| "epoch": 24.28, |
| "learning_rate": 0.000308625, |
| "loss": 0.2806, |
| "step": 777 |
| }, |
| { |
| "epoch": 24.31, |
| "learning_rate": 0.00030825, |
| "loss": 0.2927, |
| "step": 778 |
| }, |
| { |
| "epoch": 24.34, |
| "learning_rate": 0.000307875, |
| "loss": 0.4275, |
| "step": 779 |
| }, |
| { |
| "epoch": 24.38, |
| "learning_rate": 0.00030749999999999994, |
| "loss": 0.3092, |
| "step": 780 |
| }, |
| { |
| "epoch": 24.38, |
| "eval_loss": 1.0595921277999878, |
| "eval_mean_accuracy": 0.5347313224374974, |
| "eval_mean_iou": 0.4094281216720854, |
| "eval_overall_accuracy": 0.6985688460500616, |
| "eval_per_category_accuracy": [ |
| 0.7289572442338944, |
| 0.7882637927385271, |
| 0.7110830007313581, |
| 0.3521245515298298, |
| 0.5152641676475415, |
| 0.5255680684041097, |
| 0.6560224451270489, |
| 0.4452888694122324, |
| 0.3969858832506677, |
| 0.5362886801181416, |
| 0.2634269220998246, |
| 0.4975022439567922 |
| ], |
| "eval_per_category_iou": [ |
| 0.6012001790422692, |
| 0.5800720856193872, |
| 0.5888761018030578, |
| 0.2908739144142853, |
| 0.38459200476623234, |
| 0.3977994116015042, |
| 0.4489334823773417, |
| 0.40195001266995634, |
| 0.258174432124732, |
| 0.4205422401484778, |
| 0.18861287954328793, |
| 0.351510715954493 |
| ], |
| "eval_runtime": 124.0635, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 780 |
| }, |
| { |
| "epoch": 24.41, |
| "learning_rate": 0.00030712499999999994, |
| "loss": 0.2797, |
| "step": 781 |
| }, |
| { |
| "epoch": 24.44, |
| "learning_rate": 0.00030674999999999995, |
| "loss": 0.348, |
| "step": 782 |
| }, |
| { |
| "epoch": 24.47, |
| "learning_rate": 0.00030637499999999995, |
| "loss": 0.3049, |
| "step": 783 |
| }, |
| { |
| "epoch": 24.5, |
| "learning_rate": 0.00030599999999999996, |
| "loss": 0.3104, |
| "step": 784 |
| }, |
| { |
| "epoch": 24.53, |
| "learning_rate": 0.00030562499999999996, |
| "loss": 0.272, |
| "step": 785 |
| }, |
| { |
| "epoch": 24.56, |
| "learning_rate": 0.00030525, |
| "loss": 0.3088, |
| "step": 786 |
| }, |
| { |
| "epoch": 24.59, |
| "learning_rate": 0.000304875, |
| "loss": 0.3618, |
| "step": 787 |
| }, |
| { |
| "epoch": 24.62, |
| "learning_rate": 0.0003044999999999999, |
| "loss": 0.3229, |
| "step": 788 |
| }, |
| { |
| "epoch": 24.66, |
| "learning_rate": 0.000304125, |
| "loss": 0.305, |
| "step": 789 |
| }, |
| { |
| "epoch": 24.69, |
| "learning_rate": 0.00030375, |
| "loss": 0.3447, |
| "step": 790 |
| }, |
| { |
| "epoch": 24.72, |
| "learning_rate": 0.000303375, |
| "loss": 0.3447, |
| "step": 791 |
| }, |
| { |
| "epoch": 24.75, |
| "learning_rate": 0.000303, |
| "loss": 0.3794, |
| "step": 792 |
| }, |
| { |
| "epoch": 24.78, |
| "learning_rate": 0.000302625, |
| "loss": 0.366, |
| "step": 793 |
| }, |
| { |
| "epoch": 24.81, |
| "learning_rate": 0.00030225, |
| "loss": 0.3687, |
| "step": 794 |
| }, |
| { |
| "epoch": 24.84, |
| "learning_rate": 0.000301875, |
| "loss": 0.3511, |
| "step": 795 |
| }, |
| { |
| "epoch": 24.88, |
| "learning_rate": 0.00030149999999999996, |
| "loss": 0.3322, |
| "step": 796 |
| }, |
| { |
| "epoch": 24.91, |
| "learning_rate": 0.00030112499999999996, |
| "loss": 0.415, |
| "step": 797 |
| }, |
| { |
| "epoch": 24.94, |
| "learning_rate": 0.00030074999999999996, |
| "loss": 0.2909, |
| "step": 798 |
| }, |
| { |
| "epoch": 24.97, |
| "learning_rate": 0.00030037499999999997, |
| "loss": 0.3973, |
| "step": 799 |
| }, |
| { |
| "epoch": 25.0, |
| "learning_rate": 0.0003, |
| "loss": 0.3129, |
| "step": 800 |
| }, |
| { |
| "epoch": 25.0, |
| "eval_loss": 1.0445520877838135, |
| "eval_mean_accuracy": 0.5559206561600901, |
| "eval_mean_iou": 0.41632268093647634, |
| "eval_overall_accuracy": 0.7014510171455249, |
| "eval_per_category_accuracy": [ |
| 0.741826073306717, |
| 0.7631689323163794, |
| 0.7185064428286525, |
| 0.3900111727191345, |
| 0.5618599412082892, |
| 0.6393069583616112, |
| 0.6322146456973088, |
| 0.49334037635900574, |
| 0.3234474681241848, |
| 0.6506870641470229, |
| 0.27959707755736685, |
| 0.4770817212954079 |
| ], |
| "eval_per_category_iou": [ |
| 0.6156530286665254, |
| 0.5774240725503985, |
| 0.5936440463181372, |
| 0.2814546237614589, |
| 0.3967963052781128, |
| 0.4145082840946794, |
| 0.4683995464736225, |
| 0.4305084682373313, |
| 0.23923504637400103, |
| 0.4315271665216254, |
| 0.19078011666349912, |
| 0.35594146629832296 |
| ], |
| "eval_runtime": 123.7504, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 800 |
| }, |
| { |
| "epoch": 25.03, |
| "learning_rate": 0.000299625, |
| "loss": 0.3416, |
| "step": 801 |
| }, |
| { |
| "epoch": 25.06, |
| "learning_rate": 0.00029925, |
| "loss": 0.3375, |
| "step": 802 |
| }, |
| { |
| "epoch": 25.09, |
| "learning_rate": 0.000298875, |
| "loss": 0.2755, |
| "step": 803 |
| }, |
| { |
| "epoch": 25.12, |
| "learning_rate": 0.0002985, |
| "loss": 0.3397, |
| "step": 804 |
| }, |
| { |
| "epoch": 25.16, |
| "learning_rate": 0.000298125, |
| "loss": 0.2699, |
| "step": 805 |
| }, |
| { |
| "epoch": 25.19, |
| "learning_rate": 0.00029775, |
| "loss": 0.3512, |
| "step": 806 |
| }, |
| { |
| "epoch": 25.22, |
| "learning_rate": 0.00029737499999999995, |
| "loss": 0.3283, |
| "step": 807 |
| }, |
| { |
| "epoch": 25.25, |
| "learning_rate": 0.00029699999999999996, |
| "loss": 0.3389, |
| "step": 808 |
| }, |
| { |
| "epoch": 25.28, |
| "learning_rate": 0.00029662499999999996, |
| "loss": 0.3003, |
| "step": 809 |
| }, |
| { |
| "epoch": 25.31, |
| "learning_rate": 0.00029624999999999996, |
| "loss": 0.3516, |
| "step": 810 |
| }, |
| { |
| "epoch": 25.34, |
| "learning_rate": 0.00029587499999999997, |
| "loss": 0.2944, |
| "step": 811 |
| }, |
| { |
| "epoch": 25.38, |
| "learning_rate": 0.00029549999999999997, |
| "loss": 0.2966, |
| "step": 812 |
| }, |
| { |
| "epoch": 25.41, |
| "learning_rate": 0.000295125, |
| "loss": 0.3157, |
| "step": 813 |
| }, |
| { |
| "epoch": 25.44, |
| "learning_rate": 0.00029475, |
| "loss": 0.297, |
| "step": 814 |
| }, |
| { |
| "epoch": 25.47, |
| "learning_rate": 0.000294375, |
| "loss": 0.3948, |
| "step": 815 |
| }, |
| { |
| "epoch": 25.5, |
| "learning_rate": 0.000294, |
| "loss": 0.3144, |
| "step": 816 |
| }, |
| { |
| "epoch": 25.53, |
| "learning_rate": 0.000293625, |
| "loss": 0.3313, |
| "step": 817 |
| }, |
| { |
| "epoch": 25.56, |
| "learning_rate": 0.00029325, |
| "loss": 0.3419, |
| "step": 818 |
| }, |
| { |
| "epoch": 25.59, |
| "learning_rate": 0.00029287499999999995, |
| "loss": 0.3202, |
| "step": 819 |
| }, |
| { |
| "epoch": 25.62, |
| "learning_rate": 0.00029249999999999995, |
| "loss": 0.3156, |
| "step": 820 |
| }, |
| { |
| "epoch": 25.62, |
| "eval_loss": 1.0807793140411377, |
| "eval_mean_accuracy": 0.535888924465505, |
| "eval_mean_iou": 0.40688350918198, |
| "eval_overall_accuracy": 0.6983327363666735, |
| "eval_per_category_accuracy": [ |
| 0.7020573416039914, |
| 0.7963451391380512, |
| 0.7481517013136767, |
| 0.380639273842652, |
| 0.5106349788394411, |
| 0.5575163588457532, |
| 0.5494805155955532, |
| 0.48537840012600303, |
| 0.47066848264908656, |
| 0.6067730594290799, |
| 0.2852717828797752, |
| 0.3377500593229957 |
| ], |
| "eval_per_category_iou": [ |
| 0.5974668877544581, |
| 0.5835170984982666, |
| 0.5958026041361306, |
| 0.2768046387694544, |
| 0.39513924161759817, |
| 0.405876882184933, |
| 0.4471515393007436, |
| 0.42420981955989806, |
| 0.24628654403972472, |
| 0.42671514545102435, |
| 0.19208352568752884, |
| 0.2915481831839989 |
| ], |
| "eval_runtime": 123.0286, |
| "eval_samples_per_second": 0.927, |
| "eval_steps_per_second": 0.081, |
| "step": 820 |
| }, |
| { |
| "epoch": 25.66, |
| "learning_rate": 0.00029212499999999996, |
| "loss": 0.3043, |
| "step": 821 |
| }, |
| { |
| "epoch": 25.69, |
| "learning_rate": 0.00029174999999999996, |
| "loss": 0.2945, |
| "step": 822 |
| }, |
| { |
| "epoch": 25.72, |
| "learning_rate": 0.00029137499999999997, |
| "loss": 0.3044, |
| "step": 823 |
| }, |
| { |
| "epoch": 25.75, |
| "learning_rate": 0.00029099999999999997, |
| "loss": 0.3067, |
| "step": 824 |
| }, |
| { |
| "epoch": 25.78, |
| "learning_rate": 0.000290625, |
| "loss": 0.2943, |
| "step": 825 |
| }, |
| { |
| "epoch": 25.81, |
| "learning_rate": 0.00029025, |
| "loss": 0.3578, |
| "step": 826 |
| }, |
| { |
| "epoch": 25.84, |
| "learning_rate": 0.000289875, |
| "loss": 0.3079, |
| "step": 827 |
| }, |
| { |
| "epoch": 25.88, |
| "learning_rate": 0.0002895, |
| "loss": 0.3037, |
| "step": 828 |
| }, |
| { |
| "epoch": 25.91, |
| "learning_rate": 0.000289125, |
| "loss": 0.3523, |
| "step": 829 |
| }, |
| { |
| "epoch": 25.94, |
| "learning_rate": 0.00028875, |
| "loss": 0.3301, |
| "step": 830 |
| }, |
| { |
| "epoch": 25.97, |
| "learning_rate": 0.000288375, |
| "loss": 0.3635, |
| "step": 831 |
| }, |
| { |
| "epoch": 26.0, |
| "learning_rate": 0.00028799999999999995, |
| "loss": 0.3694, |
| "step": 832 |
| }, |
| { |
| "epoch": 26.03, |
| "learning_rate": 0.00028762499999999996, |
| "loss": 0.341, |
| "step": 833 |
| }, |
| { |
| "epoch": 26.06, |
| "learning_rate": 0.00028724999999999996, |
| "loss": 0.2993, |
| "step": 834 |
| }, |
| { |
| "epoch": 26.09, |
| "learning_rate": 0.00028687499999999997, |
| "loss": 0.336, |
| "step": 835 |
| }, |
| { |
| "epoch": 26.12, |
| "learning_rate": 0.00028649999999999997, |
| "loss": 0.2962, |
| "step": 836 |
| }, |
| { |
| "epoch": 26.16, |
| "learning_rate": 0.000286125, |
| "loss": 0.3283, |
| "step": 837 |
| }, |
| { |
| "epoch": 26.19, |
| "learning_rate": 0.00028575, |
| "loss": 0.2963, |
| "step": 838 |
| }, |
| { |
| "epoch": 26.22, |
| "learning_rate": 0.000285375, |
| "loss": 0.2542, |
| "step": 839 |
| }, |
| { |
| "epoch": 26.25, |
| "learning_rate": 0.000285, |
| "loss": 0.3323, |
| "step": 840 |
| }, |
| { |
| "epoch": 26.25, |
| "eval_loss": 1.058910846710205, |
| "eval_mean_accuracy": 0.540995444945956, |
| "eval_mean_iou": 0.41003933730151076, |
| "eval_overall_accuracy": 0.7011122452585321, |
| "eval_per_category_accuracy": [ |
| 0.7219966993137692, |
| 0.7921079297928585, |
| 0.7336415329243853, |
| 0.372946752624895, |
| 0.596709528025332, |
| 0.46846544409321367, |
| 0.5795466890446179, |
| 0.5022216603040144, |
| 0.442040140901307, |
| 0.5676260234770604, |
| 0.27813401976658414, |
| 0.4365089190834339 |
| ], |
| "eval_per_category_iou": [ |
| 0.607124974887669, |
| 0.5831445120530768, |
| 0.5979444309810793, |
| 0.28692776085615973, |
| 0.38274166951464034, |
| 0.3673477809240577, |
| 0.45869486386763214, |
| 0.4325816150806322, |
| 0.2529884777298741, |
| 0.4265978054403818, |
| 0.18886434065831287, |
| 0.3355138156246134 |
| ], |
| "eval_runtime": 124.9113, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 840 |
| }, |
| { |
| "epoch": 26.28, |
| "learning_rate": 0.000284625, |
| "loss": 0.314, |
| "step": 841 |
| }, |
| { |
| "epoch": 26.31, |
| "learning_rate": 0.00028425, |
| "loss": 0.2745, |
| "step": 842 |
| }, |
| { |
| "epoch": 26.34, |
| "learning_rate": 0.000283875, |
| "loss": 0.3064, |
| "step": 843 |
| }, |
| { |
| "epoch": 26.38, |
| "learning_rate": 0.00028349999999999995, |
| "loss": 0.3146, |
| "step": 844 |
| }, |
| { |
| "epoch": 26.41, |
| "learning_rate": 0.00028312499999999996, |
| "loss": 0.3282, |
| "step": 845 |
| }, |
| { |
| "epoch": 26.44, |
| "learning_rate": 0.00028274999999999996, |
| "loss": 0.2903, |
| "step": 846 |
| }, |
| { |
| "epoch": 26.47, |
| "learning_rate": 0.00028237499999999997, |
| "loss": 0.3188, |
| "step": 847 |
| }, |
| { |
| "epoch": 26.5, |
| "learning_rate": 0.00028199999999999997, |
| "loss": 0.3674, |
| "step": 848 |
| }, |
| { |
| "epoch": 26.53, |
| "learning_rate": 0.000281625, |
| "loss": 0.3132, |
| "step": 849 |
| }, |
| { |
| "epoch": 26.56, |
| "learning_rate": 0.00028125, |
| "loss": 0.3004, |
| "step": 850 |
| }, |
| { |
| "epoch": 26.59, |
| "learning_rate": 0.000280875, |
| "loss": 0.293, |
| "step": 851 |
| }, |
| { |
| "epoch": 26.62, |
| "learning_rate": 0.0002805, |
| "loss": 0.2597, |
| "step": 852 |
| }, |
| { |
| "epoch": 26.66, |
| "learning_rate": 0.000280125, |
| "loss": 0.2317, |
| "step": 853 |
| }, |
| { |
| "epoch": 26.69, |
| "learning_rate": 0.00027975, |
| "loss": 0.3386, |
| "step": 854 |
| }, |
| { |
| "epoch": 26.72, |
| "learning_rate": 0.000279375, |
| "loss": 0.3564, |
| "step": 855 |
| }, |
| { |
| "epoch": 26.75, |
| "learning_rate": 0.000279, |
| "loss": 0.3239, |
| "step": 856 |
| }, |
| { |
| "epoch": 26.78, |
| "learning_rate": 0.00027862499999999996, |
| "loss": 0.2938, |
| "step": 857 |
| }, |
| { |
| "epoch": 26.81, |
| "learning_rate": 0.00027824999999999996, |
| "loss": 0.3463, |
| "step": 858 |
| }, |
| { |
| "epoch": 26.84, |
| "learning_rate": 0.00027787499999999996, |
| "loss": 0.4822, |
| "step": 859 |
| }, |
| { |
| "epoch": 26.88, |
| "learning_rate": 0.00027749999999999997, |
| "loss": 0.3547, |
| "step": 860 |
| }, |
| { |
| "epoch": 26.88, |
| "eval_loss": 1.0882172584533691, |
| "eval_mean_accuracy": 0.5352563979463306, |
| "eval_mean_iou": 0.40511003286789854, |
| "eval_overall_accuracy": 0.6980028989022238, |
| "eval_per_category_accuracy": [ |
| 0.6688469580251329, |
| 0.7931693543314675, |
| 0.7541480961717411, |
| 0.35181574034878316, |
| 0.5413415936159739, |
| 0.5131124813547265, |
| 0.6477393493495227, |
| 0.5263872385517796, |
| 0.38461930915769765, |
| 0.6565425081661721, |
| 0.177917621888324, |
| 0.4074365243946475 |
| ], |
| "eval_per_category_iou": [ |
| 0.5885868338327909, |
| 0.5845779462406352, |
| 0.5907287320168407, |
| 0.26147434872530156, |
| 0.39272544492275857, |
| 0.3965131133357359, |
| 0.4567495599434463, |
| 0.4312363560715317, |
| 0.24811258657203367, |
| 0.4351781655388179, |
| 0.14944544141012656, |
| 0.32599186580476447 |
| ], |
| "eval_runtime": 123.4891, |
| "eval_samples_per_second": 0.923, |
| "eval_steps_per_second": 0.081, |
| "step": 860 |
| }, |
| { |
| "epoch": 26.91, |
| "learning_rate": 0.000277125, |
| "loss": 0.3587, |
| "step": 861 |
| }, |
| { |
| "epoch": 26.94, |
| "learning_rate": 0.00027675, |
| "loss": 0.3011, |
| "step": 862 |
| }, |
| { |
| "epoch": 26.97, |
| "learning_rate": 0.000276375, |
| "loss": 0.3935, |
| "step": 863 |
| }, |
| { |
| "epoch": 27.0, |
| "learning_rate": 0.000276, |
| "loss": 0.337, |
| "step": 864 |
| }, |
| { |
| "epoch": 27.03, |
| "learning_rate": 0.00027562499999999994, |
| "loss": 0.3215, |
| "step": 865 |
| }, |
| { |
| "epoch": 27.06, |
| "learning_rate": 0.00027525, |
| "loss": 0.3343, |
| "step": 866 |
| }, |
| { |
| "epoch": 27.09, |
| "learning_rate": 0.000274875, |
| "loss": 0.2881, |
| "step": 867 |
| }, |
| { |
| "epoch": 27.12, |
| "learning_rate": 0.0002745, |
| "loss": 0.303, |
| "step": 868 |
| }, |
| { |
| "epoch": 27.16, |
| "learning_rate": 0.00027412499999999995, |
| "loss": 0.3023, |
| "step": 869 |
| }, |
| { |
| "epoch": 27.19, |
| "learning_rate": 0.00027374999999999996, |
| "loss": 0.3123, |
| "step": 870 |
| }, |
| { |
| "epoch": 27.22, |
| "learning_rate": 0.00027337499999999996, |
| "loss": 0.2789, |
| "step": 871 |
| }, |
| { |
| "epoch": 27.25, |
| "learning_rate": 0.00027299999999999997, |
| "loss": 0.3179, |
| "step": 872 |
| }, |
| { |
| "epoch": 27.28, |
| "learning_rate": 0.00027262499999999997, |
| "loss": 0.3242, |
| "step": 873 |
| }, |
| { |
| "epoch": 27.31, |
| "learning_rate": 0.00027225, |
| "loss": 0.3122, |
| "step": 874 |
| }, |
| { |
| "epoch": 27.34, |
| "learning_rate": 0.000271875, |
| "loss": 0.3223, |
| "step": 875 |
| }, |
| { |
| "epoch": 27.38, |
| "learning_rate": 0.0002715, |
| "loss": 0.2767, |
| "step": 876 |
| }, |
| { |
| "epoch": 27.41, |
| "learning_rate": 0.000271125, |
| "loss": 0.2805, |
| "step": 877 |
| }, |
| { |
| "epoch": 27.44, |
| "learning_rate": 0.00027074999999999994, |
| "loss": 0.3317, |
| "step": 878 |
| }, |
| { |
| "epoch": 27.47, |
| "learning_rate": 0.000270375, |
| "loss": 0.3175, |
| "step": 879 |
| }, |
| { |
| "epoch": 27.5, |
| "learning_rate": 0.00027, |
| "loss": 0.2781, |
| "step": 880 |
| }, |
| { |
| "epoch": 27.5, |
| "eval_loss": 1.0291202068328857, |
| "eval_mean_accuracy": 0.5470684935504814, |
| "eval_mean_iou": 0.4155284278027596, |
| "eval_overall_accuracy": 0.7040024472956072, |
| "eval_per_category_accuracy": [ |
| 0.763472014136652, |
| 0.7595927182509237, |
| 0.7593774638394334, |
| 0.4153857363932242, |
| 0.5005702917503196, |
| 0.5566907044011918, |
| 0.600428403414002, |
| 0.4858101961725513, |
| 0.41753546933089625, |
| 0.5952707914680647, |
| 0.1953549445537827, |
| 0.5153331888947352 |
| ], |
| "eval_per_category_iou": [ |
| 0.6198829126480283, |
| 0.5804950384144377, |
| 0.598225011689386, |
| 0.29792671970773493, |
| 0.3953745885598434, |
| 0.41083507538940145, |
| 0.45463827930531314, |
| 0.42482310436842263, |
| 0.25560628240399996, |
| 0.42637430044256897, |
| 0.15905590693640753, |
| 0.36310391376757284 |
| ], |
| "eval_runtime": 124.066, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 880 |
| }, |
| { |
| "epoch": 27.53, |
| "learning_rate": 0.000269625, |
| "loss": 0.3119, |
| "step": 881 |
| }, |
| { |
| "epoch": 27.56, |
| "learning_rate": 0.00026924999999999996, |
| "loss": 0.3683, |
| "step": 882 |
| }, |
| { |
| "epoch": 27.59, |
| "learning_rate": 0.00026887499999999996, |
| "loss": 0.2754, |
| "step": 883 |
| }, |
| { |
| "epoch": 27.62, |
| "learning_rate": 0.00026849999999999997, |
| "loss": 0.3038, |
| "step": 884 |
| }, |
| { |
| "epoch": 27.66, |
| "learning_rate": 0.00026812499999999997, |
| "loss": 0.3114, |
| "step": 885 |
| }, |
| { |
| "epoch": 27.69, |
| "learning_rate": 0.00026775, |
| "loss": 0.3176, |
| "step": 886 |
| }, |
| { |
| "epoch": 27.72, |
| "learning_rate": 0.000267375, |
| "loss": 0.2982, |
| "step": 887 |
| }, |
| { |
| "epoch": 27.75, |
| "learning_rate": 0.000267, |
| "loss": 0.3299, |
| "step": 888 |
| }, |
| { |
| "epoch": 27.78, |
| "learning_rate": 0.000266625, |
| "loss": 0.2977, |
| "step": 889 |
| }, |
| { |
| "epoch": 27.81, |
| "learning_rate": 0.00026624999999999994, |
| "loss": 0.3098, |
| "step": 890 |
| }, |
| { |
| "epoch": 27.84, |
| "learning_rate": 0.000265875, |
| "loss": 0.2867, |
| "step": 891 |
| }, |
| { |
| "epoch": 27.88, |
| "learning_rate": 0.0002655, |
| "loss": 0.278, |
| "step": 892 |
| }, |
| { |
| "epoch": 27.91, |
| "learning_rate": 0.000265125, |
| "loss": 0.3452, |
| "step": 893 |
| }, |
| { |
| "epoch": 27.94, |
| "learning_rate": 0.00026474999999999996, |
| "loss": 0.3552, |
| "step": 894 |
| }, |
| { |
| "epoch": 27.97, |
| "learning_rate": 0.00026437499999999996, |
| "loss": 0.3034, |
| "step": 895 |
| }, |
| { |
| "epoch": 28.0, |
| "learning_rate": 0.00026399999999999997, |
| "loss": 0.2706, |
| "step": 896 |
| }, |
| { |
| "epoch": 28.03, |
| "learning_rate": 0.00026362499999999997, |
| "loss": 0.272, |
| "step": 897 |
| }, |
| { |
| "epoch": 28.06, |
| "learning_rate": 0.00026325, |
| "loss": 0.3359, |
| "step": 898 |
| }, |
| { |
| "epoch": 28.09, |
| "learning_rate": 0.000262875, |
| "loss": 0.3136, |
| "step": 899 |
| }, |
| { |
| "epoch": 28.12, |
| "learning_rate": 0.0002625, |
| "loss": 0.3065, |
| "step": 900 |
| }, |
| { |
| "epoch": 28.12, |
| "eval_loss": 1.0788742303848267, |
| "eval_mean_accuracy": 0.5399525229367704, |
| "eval_mean_iou": 0.41133326662368447, |
| "eval_overall_accuracy": 0.7011427628366571, |
| "eval_per_category_accuracy": [ |
| 0.6906467879541839, |
| 0.8000925846128125, |
| 0.7581761303307034, |
| 0.3765206348325133, |
| 0.5795359466133989, |
| 0.5740766279909559, |
| 0.5950997325037117, |
| 0.505899962606248, |
| 0.3638345918032351, |
| 0.5599759252430815, |
| 0.2104109723212728, |
| 0.46516037842912705 |
| ], |
| "eval_per_category_iou": [ |
| 0.5935301885163856, |
| 0.5857002053747803, |
| 0.5993743184524172, |
| 0.2839751390251881, |
| 0.39530082538875355, |
| 0.40564979285460584, |
| 0.4657372712296089, |
| 0.43648381311590584, |
| 0.2347614256209289, |
| 0.4166158027299286, |
| 0.16865140698216433, |
| 0.3502190101935477 |
| ], |
| "eval_runtime": 124.6664, |
| "eval_samples_per_second": 0.914, |
| "eval_steps_per_second": 0.08, |
| "step": 900 |
| }, |
| { |
| "epoch": 28.16, |
| "learning_rate": 0.000262125, |
| "loss": 0.2946, |
| "step": 901 |
| }, |
| { |
| "epoch": 28.19, |
| "learning_rate": 0.00026175, |
| "loss": 0.3188, |
| "step": 902 |
| }, |
| { |
| "epoch": 28.22, |
| "learning_rate": 0.00026137499999999994, |
| "loss": 0.2962, |
| "step": 903 |
| }, |
| { |
| "epoch": 28.25, |
| "learning_rate": 0.000261, |
| "loss": 0.3317, |
| "step": 904 |
| }, |
| { |
| "epoch": 28.28, |
| "learning_rate": 0.000260625, |
| "loss": 0.2952, |
| "step": 905 |
| }, |
| { |
| "epoch": 28.31, |
| "learning_rate": 0.00026025, |
| "loss": 0.2838, |
| "step": 906 |
| }, |
| { |
| "epoch": 28.34, |
| "learning_rate": 0.00025987499999999996, |
| "loss": 0.3201, |
| "step": 907 |
| }, |
| { |
| "epoch": 28.38, |
| "learning_rate": 0.00025949999999999997, |
| "loss": 0.2928, |
| "step": 908 |
| }, |
| { |
| "epoch": 28.41, |
| "learning_rate": 0.00025912499999999997, |
| "loss": 0.3303, |
| "step": 909 |
| }, |
| { |
| "epoch": 28.44, |
| "learning_rate": 0.00025875, |
| "loss": 0.3442, |
| "step": 910 |
| }, |
| { |
| "epoch": 28.47, |
| "learning_rate": 0.000258375, |
| "loss": 0.3448, |
| "step": 911 |
| }, |
| { |
| "epoch": 28.5, |
| "learning_rate": 0.000258, |
| "loss": 0.3323, |
| "step": 912 |
| }, |
| { |
| "epoch": 28.53, |
| "learning_rate": 0.000257625, |
| "loss": 0.2975, |
| "step": 913 |
| }, |
| { |
| "epoch": 28.56, |
| "learning_rate": 0.00025725, |
| "loss": 0.3361, |
| "step": 914 |
| }, |
| { |
| "epoch": 28.59, |
| "learning_rate": 0.00025687499999999994, |
| "loss": 0.3285, |
| "step": 915 |
| }, |
| { |
| "epoch": 28.62, |
| "learning_rate": 0.00025649999999999995, |
| "loss": 0.2993, |
| "step": 916 |
| }, |
| { |
| "epoch": 28.66, |
| "learning_rate": 0.000256125, |
| "loss": 0.3384, |
| "step": 917 |
| }, |
| { |
| "epoch": 28.69, |
| "learning_rate": 0.00025575, |
| "loss": 0.2851, |
| "step": 918 |
| }, |
| { |
| "epoch": 28.72, |
| "learning_rate": 0.00025537499999999996, |
| "loss": 0.2692, |
| "step": 919 |
| }, |
| { |
| "epoch": 28.75, |
| "learning_rate": 0.00025499999999999996, |
| "loss": 0.3251, |
| "step": 920 |
| }, |
| { |
| "epoch": 28.75, |
| "eval_loss": 1.0922932624816895, |
| "eval_mean_accuracy": 0.5374819353041405, |
| "eval_mean_iou": 0.40751518063238884, |
| "eval_overall_accuracy": 0.6977377105177495, |
| "eval_per_category_accuracy": [ |
| 0.6692330023396136, |
| 0.7984075155551366, |
| 0.7547019651362172, |
| 0.33764859369470024, |
| 0.5991163010249044, |
| 0.5182188365349369, |
| 0.604408571846921, |
| 0.4938139591197361, |
| 0.40590756235415204, |
| 0.5979531241117348, |
| 0.1759862631728765, |
| 0.4943875287587566 |
| ], |
| "eval_per_category_iou": [ |
| 0.5827423705902194, |
| 0.5837738180979003, |
| 0.5973455144999549, |
| 0.2732633730873024, |
| 0.39792618946059666, |
| 0.3975131330959586, |
| 0.4599646659833107, |
| 0.42106907666332893, |
| 0.2492450648283978, |
| 0.42439733492161086, |
| 0.14455271798430203, |
| 0.35838890837578274 |
| ], |
| "eval_runtime": 123.476, |
| "eval_samples_per_second": 0.923, |
| "eval_steps_per_second": 0.081, |
| "step": 920 |
| }, |
| { |
| "epoch": 28.78, |
| "learning_rate": 0.00025462499999999997, |
| "loss": 0.2983, |
| "step": 921 |
| }, |
| { |
| "epoch": 28.81, |
| "learning_rate": 0.00025425, |
| "loss": 0.2995, |
| "step": 922 |
| }, |
| { |
| "epoch": 28.84, |
| "learning_rate": 0.000253875, |
| "loss": 0.2764, |
| "step": 923 |
| }, |
| { |
| "epoch": 28.88, |
| "learning_rate": 0.0002535, |
| "loss": 0.3144, |
| "step": 924 |
| }, |
| { |
| "epoch": 28.91, |
| "learning_rate": 0.000253125, |
| "loss": 0.3135, |
| "step": 925 |
| }, |
| { |
| "epoch": 28.94, |
| "learning_rate": 0.00025275, |
| "loss": 0.3024, |
| "step": 926 |
| }, |
| { |
| "epoch": 28.97, |
| "learning_rate": 0.000252375, |
| "loss": 0.2667, |
| "step": 927 |
| }, |
| { |
| "epoch": 29.0, |
| "learning_rate": 0.00025199999999999995, |
| "loss": 0.3208, |
| "step": 928 |
| }, |
| { |
| "epoch": 29.03, |
| "learning_rate": 0.00025162499999999995, |
| "loss": 0.2899, |
| "step": 929 |
| }, |
| { |
| "epoch": 29.06, |
| "learning_rate": 0.00025125, |
| "loss": 0.3329, |
| "step": 930 |
| }, |
| { |
| "epoch": 29.09, |
| "learning_rate": 0.000250875, |
| "loss": 0.281, |
| "step": 931 |
| }, |
| { |
| "epoch": 29.12, |
| "learning_rate": 0.00025049999999999996, |
| "loss": 0.2907, |
| "step": 932 |
| }, |
| { |
| "epoch": 29.16, |
| "learning_rate": 0.00025012499999999997, |
| "loss": 0.3136, |
| "step": 933 |
| }, |
| { |
| "epoch": 29.19, |
| "learning_rate": 0.00024974999999999997, |
| "loss": 0.3556, |
| "step": 934 |
| }, |
| { |
| "epoch": 29.22, |
| "learning_rate": 0.000249375, |
| "loss": 0.2707, |
| "step": 935 |
| }, |
| { |
| "epoch": 29.25, |
| "learning_rate": 0.000249, |
| "loss": 0.272, |
| "step": 936 |
| }, |
| { |
| "epoch": 29.28, |
| "learning_rate": 0.000248625, |
| "loss": 0.2922, |
| "step": 937 |
| }, |
| { |
| "epoch": 29.31, |
| "learning_rate": 0.00024825, |
| "loss": 0.305, |
| "step": 938 |
| }, |
| { |
| "epoch": 29.34, |
| "learning_rate": 0.000247875, |
| "loss": 0.2655, |
| "step": 939 |
| }, |
| { |
| "epoch": 29.38, |
| "learning_rate": 0.00024749999999999994, |
| "loss": 0.2869, |
| "step": 940 |
| }, |
| { |
| "epoch": 29.38, |
| "eval_loss": 1.0787373781204224, |
| "eval_mean_accuracy": 0.5396699389625045, |
| "eval_mean_iou": 0.40855088577964166, |
| "eval_overall_accuracy": 0.7003560986435204, |
| "eval_per_category_accuracy": [ |
| 0.690306264270167, |
| 0.7930514811516904, |
| 0.7787046657111166, |
| 0.3804553750494445, |
| 0.6042600692453713, |
| 0.5699610581442192, |
| 0.6010014972864635, |
| 0.4874655929266879, |
| 0.35244847075942964, |
| 0.5914791215634406, |
| 0.23453306112704422, |
| 0.39237261031497933 |
| ], |
| "eval_per_category_iou": [ |
| 0.5928782781024985, |
| 0.584810920773929, |
| 0.5974365766728258, |
| 0.2742091179633381, |
| 0.4010355741641591, |
| 0.4010431696012462, |
| 0.4614820599801976, |
| 0.4211441763677834, |
| 0.24073975869660408, |
| 0.4324020433576768, |
| 0.1746331279898445, |
| 0.32079582568559706 |
| ], |
| "eval_runtime": 124.2274, |
| "eval_samples_per_second": 0.918, |
| "eval_steps_per_second": 0.08, |
| "step": 940 |
| }, |
| { |
| "epoch": 29.41, |
| "learning_rate": 0.00024712499999999995, |
| "loss": 0.3059, |
| "step": 941 |
| }, |
| { |
| "epoch": 29.44, |
| "learning_rate": 0.00024675, |
| "loss": 0.3455, |
| "step": 942 |
| }, |
| { |
| "epoch": 29.47, |
| "learning_rate": 0.000246375, |
| "loss": 0.3309, |
| "step": 943 |
| }, |
| { |
| "epoch": 29.5, |
| "learning_rate": 0.00024599999999999996, |
| "loss": 0.3186, |
| "step": 944 |
| }, |
| { |
| "epoch": 29.53, |
| "learning_rate": 0.00024562499999999997, |
| "loss": 0.2922, |
| "step": 945 |
| }, |
| { |
| "epoch": 29.56, |
| "learning_rate": 0.00024524999999999997, |
| "loss": 0.3072, |
| "step": 946 |
| }, |
| { |
| "epoch": 29.59, |
| "learning_rate": 0.000244875, |
| "loss": 0.2966, |
| "step": 947 |
| }, |
| { |
| "epoch": 29.62, |
| "learning_rate": 0.0002445, |
| "loss": 0.3389, |
| "step": 948 |
| }, |
| { |
| "epoch": 29.66, |
| "learning_rate": 0.00024412499999999996, |
| "loss": 0.3114, |
| "step": 949 |
| }, |
| { |
| "epoch": 29.69, |
| "learning_rate": 0.00024375, |
| "loss": 0.2737, |
| "step": 950 |
| }, |
| { |
| "epoch": 29.72, |
| "learning_rate": 0.000243375, |
| "loss": 0.2871, |
| "step": 951 |
| }, |
| { |
| "epoch": 29.75, |
| "learning_rate": 0.000243, |
| "loss": 0.3111, |
| "step": 952 |
| }, |
| { |
| "epoch": 29.78, |
| "learning_rate": 0.00024262499999999998, |
| "loss": 0.3052, |
| "step": 953 |
| }, |
| { |
| "epoch": 29.81, |
| "learning_rate": 0.00024224999999999998, |
| "loss": 0.3036, |
| "step": 954 |
| }, |
| { |
| "epoch": 29.84, |
| "learning_rate": 0.00024187499999999998, |
| "loss": 0.3226, |
| "step": 955 |
| }, |
| { |
| "epoch": 29.88, |
| "learning_rate": 0.0002415, |
| "loss": 0.2767, |
| "step": 956 |
| }, |
| { |
| "epoch": 29.91, |
| "learning_rate": 0.00024112499999999997, |
| "loss": 0.3003, |
| "step": 957 |
| }, |
| { |
| "epoch": 29.94, |
| "learning_rate": 0.00024074999999999997, |
| "loss": 0.2928, |
| "step": 958 |
| }, |
| { |
| "epoch": 29.97, |
| "learning_rate": 0.00024037499999999997, |
| "loss": 0.3199, |
| "step": 959 |
| }, |
| { |
| "epoch": 30.0, |
| "learning_rate": 0.00023999999999999998, |
| "loss": 0.3475, |
| "step": 960 |
| }, |
| { |
| "epoch": 30.0, |
| "eval_loss": 1.0594263076782227, |
| "eval_mean_accuracy": 0.5531448205807578, |
| "eval_mean_iou": 0.41451857889653415, |
| "eval_overall_accuracy": 0.7049752285605982, |
| "eval_per_category_accuracy": [ |
| 0.717755791802733, |
| 0.7775708843867838, |
| 0.76853698980065, |
| 0.3782416499538518, |
| 0.5580390877763078, |
| 0.6247935863888596, |
| 0.5663253819182585, |
| 0.5464909338902038, |
| 0.4385442268617517, |
| 0.639225477101078, |
| 0.18067845479059602, |
| 0.4415353822980181 |
| ], |
| "eval_per_category_iou": [ |
| 0.6106760630387156, |
| 0.5874425052470347, |
| 0.6069341665889207, |
| 0.27832245677461526, |
| 0.39541724147336177, |
| 0.41014871181546586, |
| 0.4498306417978669, |
| 0.44912529729853484, |
| 0.2559017611440317, |
| 0.4350634690747998, |
| 0.15162529057447055, |
| 0.34373534193059224 |
| ], |
| "eval_runtime": 124.0206, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 960 |
| }, |
| { |
| "epoch": 30.03, |
| "learning_rate": 0.00023962499999999996, |
| "loss": 0.3113, |
| "step": 961 |
| }, |
| { |
| "epoch": 30.06, |
| "learning_rate": 0.00023925, |
| "loss": 0.3048, |
| "step": 962 |
| }, |
| { |
| "epoch": 30.09, |
| "learning_rate": 0.000238875, |
| "loss": 0.3683, |
| "step": 963 |
| }, |
| { |
| "epoch": 30.12, |
| "learning_rate": 0.0002385, |
| "loss": 0.2781, |
| "step": 964 |
| }, |
| { |
| "epoch": 30.16, |
| "learning_rate": 0.00023812499999999997, |
| "loss": 0.3513, |
| "step": 965 |
| }, |
| { |
| "epoch": 30.19, |
| "learning_rate": 0.00023774999999999998, |
| "loss": 0.355, |
| "step": 966 |
| }, |
| { |
| "epoch": 30.22, |
| "learning_rate": 0.00023737499999999998, |
| "loss": 0.2747, |
| "step": 967 |
| }, |
| { |
| "epoch": 30.25, |
| "learning_rate": 0.000237, |
| "loss": 0.3135, |
| "step": 968 |
| }, |
| { |
| "epoch": 30.28, |
| "learning_rate": 0.00023662499999999996, |
| "loss": 0.3054, |
| "step": 969 |
| }, |
| { |
| "epoch": 30.31, |
| "learning_rate": 0.00023624999999999997, |
| "loss": 0.2638, |
| "step": 970 |
| }, |
| { |
| "epoch": 30.34, |
| "learning_rate": 0.00023587499999999997, |
| "loss": 0.3067, |
| "step": 971 |
| }, |
| { |
| "epoch": 30.38, |
| "learning_rate": 0.00023549999999999998, |
| "loss": 0.3627, |
| "step": 972 |
| }, |
| { |
| "epoch": 30.41, |
| "learning_rate": 0.00023512499999999996, |
| "loss": 0.2484, |
| "step": 973 |
| }, |
| { |
| "epoch": 30.44, |
| "learning_rate": 0.00023474999999999996, |
| "loss": 0.2995, |
| "step": 974 |
| }, |
| { |
| "epoch": 30.47, |
| "learning_rate": 0.000234375, |
| "loss": 0.3064, |
| "step": 975 |
| }, |
| { |
| "epoch": 30.5, |
| "learning_rate": 0.000234, |
| "loss": 0.2741, |
| "step": 976 |
| }, |
| { |
| "epoch": 30.53, |
| "learning_rate": 0.000233625, |
| "loss": 0.2736, |
| "step": 977 |
| }, |
| { |
| "epoch": 30.56, |
| "learning_rate": 0.00023324999999999998, |
| "loss": 0.305, |
| "step": 978 |
| }, |
| { |
| "epoch": 30.59, |
| "learning_rate": 0.00023287499999999998, |
| "loss": 0.2436, |
| "step": 979 |
| }, |
| { |
| "epoch": 30.62, |
| "learning_rate": 0.00023249999999999999, |
| "loss": 0.2797, |
| "step": 980 |
| }, |
| { |
| "epoch": 30.62, |
| "eval_loss": 1.0813096761703491, |
| "eval_mean_accuracy": 0.5290516443036589, |
| "eval_mean_iou": 0.40902801888421064, |
| "eval_overall_accuracy": 0.7042823925352933, |
| "eval_per_category_accuracy": [ |
| 0.7427777481648945, |
| 0.798221731901414, |
| 0.7280928559464154, |
| 0.3263509621723652, |
| 0.5745886910031665, |
| 0.5512758629449914, |
| 0.5790239770510541, |
| 0.5382396650205559, |
| 0.3735614846099926, |
| 0.5269847462214993, |
| 0.19486521809661692, |
| 0.4146367885109412 |
| ], |
| "eval_per_category_iou": [ |
| 0.6132746633068886, |
| 0.5857632393763731, |
| 0.5958261576201113, |
| 0.26939667518302535, |
| 0.39379335432668544, |
| 0.40216307808394175, |
| 0.4569927313146199, |
| 0.44616917738321193, |
| 0.2505415312838457, |
| 0.40902062064331224, |
| 0.1599486476029686, |
| 0.3254463504855438 |
| ], |
| "eval_runtime": 123.8542, |
| "eval_samples_per_second": 0.92, |
| "eval_steps_per_second": 0.081, |
| "step": 980 |
| }, |
| { |
| "epoch": 30.66, |
| "learning_rate": 0.000232125, |
| "loss": 0.3114, |
| "step": 981 |
| }, |
| { |
| "epoch": 30.69, |
| "learning_rate": 0.00023174999999999997, |
| "loss": 0.3217, |
| "step": 982 |
| }, |
| { |
| "epoch": 30.72, |
| "learning_rate": 0.00023137499999999997, |
| "loss": 0.3091, |
| "step": 983 |
| }, |
| { |
| "epoch": 30.75, |
| "learning_rate": 0.00023099999999999998, |
| "loss": 0.2643, |
| "step": 984 |
| }, |
| { |
| "epoch": 30.78, |
| "learning_rate": 0.00023062499999999998, |
| "loss": 0.2828, |
| "step": 985 |
| }, |
| { |
| "epoch": 30.81, |
| "learning_rate": 0.00023024999999999996, |
| "loss": 0.2724, |
| "step": 986 |
| }, |
| { |
| "epoch": 30.84, |
| "learning_rate": 0.00022987499999999996, |
| "loss": 0.2631, |
| "step": 987 |
| }, |
| { |
| "epoch": 30.88, |
| "learning_rate": 0.0002295, |
| "loss": 0.2717, |
| "step": 988 |
| }, |
| { |
| "epoch": 30.91, |
| "learning_rate": 0.000229125, |
| "loss": 0.2975, |
| "step": 989 |
| }, |
| { |
| "epoch": 30.94, |
| "learning_rate": 0.00022874999999999998, |
| "loss": 0.262, |
| "step": 990 |
| }, |
| { |
| "epoch": 30.97, |
| "learning_rate": 0.00022837499999999998, |
| "loss": 0.2993, |
| "step": 991 |
| }, |
| { |
| "epoch": 31.0, |
| "learning_rate": 0.00022799999999999999, |
| "loss": 0.3107, |
| "step": 992 |
| }, |
| { |
| "epoch": 31.03, |
| "learning_rate": 0.000227625, |
| "loss": 0.2732, |
| "step": 993 |
| }, |
| { |
| "epoch": 31.06, |
| "learning_rate": 0.00022724999999999997, |
| "loss": 0.3334, |
| "step": 994 |
| }, |
| { |
| "epoch": 31.09, |
| "learning_rate": 0.00022687499999999997, |
| "loss": 0.2656, |
| "step": 995 |
| }, |
| { |
| "epoch": 31.12, |
| "learning_rate": 0.00022649999999999998, |
| "loss": 0.324, |
| "step": 996 |
| }, |
| { |
| "epoch": 31.16, |
| "learning_rate": 0.00022612499999999998, |
| "loss": 0.2526, |
| "step": 997 |
| }, |
| { |
| "epoch": 31.19, |
| "learning_rate": 0.00022574999999999996, |
| "loss": 0.3622, |
| "step": 998 |
| }, |
| { |
| "epoch": 31.22, |
| "learning_rate": 0.00022537499999999996, |
| "loss": 0.2805, |
| "step": 999 |
| }, |
| { |
| "epoch": 31.25, |
| "learning_rate": 0.000225, |
| "loss": 0.3178, |
| "step": 1000 |
| }, |
| { |
| "epoch": 31.25, |
| "eval_loss": 1.116824984550476, |
| "eval_mean_accuracy": 0.5356298327841512, |
| "eval_mean_iou": 0.4094583385141765, |
| "eval_overall_accuracy": 0.7037695834511205, |
| "eval_per_category_accuracy": [ |
| 0.7205846787249348, |
| 0.7780062390322563, |
| 0.7833884396332765, |
| 0.36228756219595976, |
| 0.5306164620842087, |
| 0.501754742522837, |
| 0.5945455318358369, |
| 0.5187756278453806, |
| 0.3699768417877075, |
| 0.6591387850279629, |
| 0.2285033041231907, |
| 0.37997977859626314 |
| ], |
| "eval_per_category_iou": [ |
| 0.6117729617771942, |
| 0.585473724108078, |
| 0.5970538087139583, |
| 0.2693620413438692, |
| 0.39267177403637127, |
| 0.389500478237045, |
| 0.45852482803421707, |
| 0.43402171457666255, |
| 0.24465554013735563, |
| 0.43876413485864185, |
| 0.17646641689614825, |
| 0.315232639450577 |
| ], |
| "eval_runtime": 124.9093, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 1000 |
| }, |
| { |
| "epoch": 31.28, |
| "learning_rate": 0.000224625, |
| "loss": 0.3342, |
| "step": 1001 |
| }, |
| { |
| "epoch": 31.31, |
| "learning_rate": 0.00022425, |
| "loss": 0.3054, |
| "step": 1002 |
| }, |
| { |
| "epoch": 31.34, |
| "learning_rate": 0.00022387499999999998, |
| "loss": 0.26, |
| "step": 1003 |
| }, |
| { |
| "epoch": 31.38, |
| "learning_rate": 0.00022349999999999998, |
| "loss": 0.2806, |
| "step": 1004 |
| }, |
| { |
| "epoch": 31.41, |
| "learning_rate": 0.000223125, |
| "loss": 0.3078, |
| "step": 1005 |
| }, |
| { |
| "epoch": 31.44, |
| "learning_rate": 0.00022275, |
| "loss": 0.3331, |
| "step": 1006 |
| }, |
| { |
| "epoch": 31.47, |
| "learning_rate": 0.00022237499999999997, |
| "loss": 0.2443, |
| "step": 1007 |
| }, |
| { |
| "epoch": 31.5, |
| "learning_rate": 0.00022199999999999998, |
| "loss": 0.2813, |
| "step": 1008 |
| }, |
| { |
| "epoch": 31.53, |
| "learning_rate": 0.00022162499999999998, |
| "loss": 0.2939, |
| "step": 1009 |
| }, |
| { |
| "epoch": 31.56, |
| "learning_rate": 0.00022124999999999998, |
| "loss": 0.2589, |
| "step": 1010 |
| }, |
| { |
| "epoch": 31.59, |
| "learning_rate": 0.00022087499999999996, |
| "loss": 0.3043, |
| "step": 1011 |
| }, |
| { |
| "epoch": 31.62, |
| "learning_rate": 0.00022049999999999997, |
| "loss": 0.2508, |
| "step": 1012 |
| }, |
| { |
| "epoch": 31.66, |
| "learning_rate": 0.000220125, |
| "loss": 0.3147, |
| "step": 1013 |
| }, |
| { |
| "epoch": 31.69, |
| "learning_rate": 0.00021975, |
| "loss": 0.2875, |
| "step": 1014 |
| }, |
| { |
| "epoch": 31.72, |
| "learning_rate": 0.00021937499999999998, |
| "loss": 0.3019, |
| "step": 1015 |
| }, |
| { |
| "epoch": 31.75, |
| "learning_rate": 0.00021899999999999998, |
| "loss": 0.2683, |
| "step": 1016 |
| }, |
| { |
| "epoch": 31.78, |
| "learning_rate": 0.000218625, |
| "loss": 0.3312, |
| "step": 1017 |
| }, |
| { |
| "epoch": 31.81, |
| "learning_rate": 0.00021825, |
| "loss": 0.33, |
| "step": 1018 |
| }, |
| { |
| "epoch": 31.84, |
| "learning_rate": 0.00021787499999999997, |
| "loss": 0.2915, |
| "step": 1019 |
| }, |
| { |
| "epoch": 31.88, |
| "learning_rate": 0.00021749999999999997, |
| "loss": 0.2949, |
| "step": 1020 |
| }, |
| { |
| "epoch": 31.88, |
| "eval_loss": 1.0969414710998535, |
| "eval_mean_accuracy": 0.5458832025922241, |
| "eval_mean_iou": 0.41571726815461857, |
| "eval_overall_accuracy": 0.6998190963477419, |
| "eval_per_category_accuracy": [ |
| 0.6632445431410028, |
| 0.805647661381561, |
| 0.7361363687597359, |
| 0.3737795018771556, |
| 0.5204788424799486, |
| 0.5500310300901143, |
| 0.622238246458705, |
| 0.5244157628057529, |
| 0.4217988873410646, |
| 0.6205477215819034, |
| 0.2583949827524463, |
| 0.4538848824372982 |
| ], |
| "eval_per_category_iou": [ |
| 0.5818945593007311, |
| 0.5864245229640183, |
| 0.5983254971231151, |
| 0.280337058183654, |
| 0.400208117193658, |
| 0.40762835633423616, |
| 0.4651237235883636, |
| 0.43865075211150467, |
| 0.25856139935792055, |
| 0.43210371529283415, |
| 0.19739244904496575, |
| 0.34195706736042075 |
| ], |
| "eval_runtime": 124.1268, |
| "eval_samples_per_second": 0.918, |
| "eval_steps_per_second": 0.081, |
| "step": 1020 |
| }, |
| { |
| "epoch": 31.91, |
| "learning_rate": 0.00021712499999999998, |
| "loss": 0.3011, |
| "step": 1021 |
| }, |
| { |
| "epoch": 31.94, |
| "learning_rate": 0.00021674999999999998, |
| "loss": 0.2443, |
| "step": 1022 |
| }, |
| { |
| "epoch": 31.97, |
| "learning_rate": 0.00021637499999999996, |
| "loss": 0.2881, |
| "step": 1023 |
| }, |
| { |
| "epoch": 32.0, |
| "learning_rate": 0.00021599999999999996, |
| "loss": 0.3192, |
| "step": 1024 |
| }, |
| { |
| "epoch": 32.03, |
| "learning_rate": 0.00021562499999999997, |
| "loss": 0.3189, |
| "step": 1025 |
| }, |
| { |
| "epoch": 32.06, |
| "learning_rate": 0.00021525, |
| "loss": 0.3017, |
| "step": 1026 |
| }, |
| { |
| "epoch": 32.09, |
| "learning_rate": 0.000214875, |
| "loss": 0.3095, |
| "step": 1027 |
| }, |
| { |
| "epoch": 32.12, |
| "learning_rate": 0.00021449999999999998, |
| "loss": 0.2525, |
| "step": 1028 |
| }, |
| { |
| "epoch": 32.16, |
| "learning_rate": 0.000214125, |
| "loss": 0.3975, |
| "step": 1029 |
| }, |
| { |
| "epoch": 32.19, |
| "learning_rate": 0.00021375, |
| "loss": 0.2452, |
| "step": 1030 |
| }, |
| { |
| "epoch": 32.22, |
| "learning_rate": 0.000213375, |
| "loss": 0.3474, |
| "step": 1031 |
| }, |
| { |
| "epoch": 32.25, |
| "learning_rate": 0.00021299999999999997, |
| "loss": 0.3063, |
| "step": 1032 |
| }, |
| { |
| "epoch": 32.28, |
| "learning_rate": 0.00021262499999999998, |
| "loss": 0.2932, |
| "step": 1033 |
| }, |
| { |
| "epoch": 32.31, |
| "learning_rate": 0.00021224999999999998, |
| "loss": 0.3601, |
| "step": 1034 |
| }, |
| { |
| "epoch": 32.34, |
| "learning_rate": 0.000211875, |
| "loss": 0.2732, |
| "step": 1035 |
| }, |
| { |
| "epoch": 32.38, |
| "learning_rate": 0.00021149999999999996, |
| "loss": 0.3098, |
| "step": 1036 |
| }, |
| { |
| "epoch": 32.41, |
| "learning_rate": 0.00021112499999999997, |
| "loss": 0.2685, |
| "step": 1037 |
| }, |
| { |
| "epoch": 32.44, |
| "learning_rate": 0.00021074999999999997, |
| "loss": 0.2966, |
| "step": 1038 |
| }, |
| { |
| "epoch": 32.47, |
| "learning_rate": 0.000210375, |
| "loss": 0.3313, |
| "step": 1039 |
| }, |
| { |
| "epoch": 32.5, |
| "learning_rate": 0.00020999999999999998, |
| "loss": 0.3418, |
| "step": 1040 |
| }, |
| { |
| "epoch": 32.5, |
| "eval_loss": 1.0765517950057983, |
| "eval_mean_accuracy": 0.5543500439241925, |
| "eval_mean_iou": 0.41639279248161637, |
| "eval_overall_accuracy": 0.7043008302387438, |
| "eval_per_category_accuracy": [ |
| 0.7614553033663681, |
| 0.7559075240439458, |
| 0.760446108492686, |
| 0.3855177965454785, |
| 0.6086765382176509, |
| 0.5261814117057839, |
| 0.633948097218133, |
| 0.5224303581549987, |
| 0.3685704905814397, |
| 0.5955451268696923, |
| 0.2588418581446101, |
| 0.47467991374952284 |
| ], |
| "eval_per_category_iou": [ |
| 0.623667834884409, |
| 0.5801914385379658, |
| 0.5951955749457164, |
| 0.27970515747843183, |
| 0.39661137769211985, |
| 0.4046224303551594, |
| 0.4698392120873024, |
| 0.43922041281800445, |
| 0.24345843553383092, |
| 0.42694784496731547, |
| 0.18333990955170426, |
| 0.35391388092743614 |
| ], |
| "eval_runtime": 125.7869, |
| "eval_samples_per_second": 0.906, |
| "eval_steps_per_second": 0.079, |
| "step": 1040 |
| }, |
| { |
| "epoch": 32.53, |
| "learning_rate": 0.00020962499999999999, |
| "loss": 0.2642, |
| "step": 1041 |
| }, |
| { |
| "epoch": 32.56, |
| "learning_rate": 0.00020925, |
| "loss": 0.2919, |
| "step": 1042 |
| }, |
| { |
| "epoch": 32.59, |
| "learning_rate": 0.000208875, |
| "loss": 0.3214, |
| "step": 1043 |
| }, |
| { |
| "epoch": 32.62, |
| "learning_rate": 0.00020849999999999997, |
| "loss": 0.2805, |
| "step": 1044 |
| }, |
| { |
| "epoch": 32.66, |
| "learning_rate": 0.00020812499999999998, |
| "loss": 0.3009, |
| "step": 1045 |
| }, |
| { |
| "epoch": 32.69, |
| "learning_rate": 0.00020774999999999998, |
| "loss": 0.3019, |
| "step": 1046 |
| }, |
| { |
| "epoch": 32.72, |
| "learning_rate": 0.00020737499999999999, |
| "loss": 0.2419, |
| "step": 1047 |
| }, |
| { |
| "epoch": 32.75, |
| "learning_rate": 0.00020699999999999996, |
| "loss": 0.2712, |
| "step": 1048 |
| }, |
| { |
| "epoch": 32.78, |
| "learning_rate": 0.00020662499999999997, |
| "loss": 0.254, |
| "step": 1049 |
| }, |
| { |
| "epoch": 32.81, |
| "learning_rate": 0.00020624999999999997, |
| "loss": 0.2614, |
| "step": 1050 |
| }, |
| { |
| "epoch": 32.84, |
| "learning_rate": 0.000205875, |
| "loss": 0.2921, |
| "step": 1051 |
| }, |
| { |
| "epoch": 32.88, |
| "learning_rate": 0.0002055, |
| "loss": 0.2337, |
| "step": 1052 |
| }, |
| { |
| "epoch": 32.91, |
| "learning_rate": 0.00020512499999999999, |
| "loss": 0.2671, |
| "step": 1053 |
| }, |
| { |
| "epoch": 32.94, |
| "learning_rate": 0.00020475, |
| "loss": 0.2414, |
| "step": 1054 |
| }, |
| { |
| "epoch": 32.97, |
| "learning_rate": 0.000204375, |
| "loss": 0.2938, |
| "step": 1055 |
| }, |
| { |
| "epoch": 33.0, |
| "learning_rate": 0.000204, |
| "loss": 0.3263, |
| "step": 1056 |
| }, |
| { |
| "epoch": 33.03, |
| "learning_rate": 0.00020362499999999998, |
| "loss": 0.2565, |
| "step": 1057 |
| }, |
| { |
| "epoch": 33.06, |
| "learning_rate": 0.00020324999999999998, |
| "loss": 0.3171, |
| "step": 1058 |
| }, |
| { |
| "epoch": 33.09, |
| "learning_rate": 0.00020287499999999998, |
| "loss": 0.247, |
| "step": 1059 |
| }, |
| { |
| "epoch": 33.12, |
| "learning_rate": 0.0002025, |
| "loss": 0.2502, |
| "step": 1060 |
| }, |
| { |
| "epoch": 33.12, |
| "eval_loss": 1.118545413017273, |
| "eval_mean_accuracy": 0.5417944713088252, |
| "eval_mean_iou": 0.4116440690273877, |
| "eval_overall_accuracy": 0.6949641244453296, |
| "eval_per_category_accuracy": [ |
| 0.6270258332514697, |
| 0.8188767833977647, |
| 0.7467714518643355, |
| 0.3557851784512252, |
| 0.5531424797460328, |
| 0.6017968418263839, |
| 0.6204410303724008, |
| 0.5547207736756559, |
| 0.3600059448284429, |
| 0.5402091729204881, |
| 0.2476026359526557, |
| 0.4751555294190473 |
| ], |
| "eval_per_category_iou": [ |
| 0.5571290028934748, |
| 0.5839884812884647, |
| 0.5965186486798824, |
| 0.28500273779181107, |
| 0.39121638370398604, |
| 0.4169573056197462, |
| 0.4680129261815746, |
| 0.45158355648207105, |
| 0.23525617330927065, |
| 0.4172661464751715, |
| 0.18655563345202295, |
| 0.35024183245117724 |
| ], |
| "eval_runtime": 124.8115, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 1060 |
| }, |
| { |
| "epoch": 33.16, |
| "learning_rate": 0.00020212499999999997, |
| "loss": 0.3372, |
| "step": 1061 |
| }, |
| { |
| "epoch": 33.19, |
| "learning_rate": 0.00020174999999999997, |
| "loss": 0.292, |
| "step": 1062 |
| }, |
| { |
| "epoch": 33.22, |
| "learning_rate": 0.00020137499999999998, |
| "loss": 0.2222, |
| "step": 1063 |
| }, |
| { |
| "epoch": 33.25, |
| "learning_rate": 0.000201, |
| "loss": 0.3246, |
| "step": 1064 |
| }, |
| { |
| "epoch": 33.28, |
| "learning_rate": 0.00020062499999999996, |
| "loss": 0.2997, |
| "step": 1065 |
| }, |
| { |
| "epoch": 33.31, |
| "learning_rate": 0.00020025, |
| "loss": 0.2881, |
| "step": 1066 |
| }, |
| { |
| "epoch": 33.34, |
| "learning_rate": 0.000199875, |
| "loss": 0.2869, |
| "step": 1067 |
| }, |
| { |
| "epoch": 33.38, |
| "learning_rate": 0.0001995, |
| "loss": 0.2806, |
| "step": 1068 |
| }, |
| { |
| "epoch": 33.41, |
| "learning_rate": 0.00019912499999999997, |
| "loss": 0.2892, |
| "step": 1069 |
| }, |
| { |
| "epoch": 33.44, |
| "learning_rate": 0.00019874999999999998, |
| "loss": 0.2847, |
| "step": 1070 |
| }, |
| { |
| "epoch": 33.47, |
| "learning_rate": 0.00019837499999999998, |
| "loss": 0.265, |
| "step": 1071 |
| }, |
| { |
| "epoch": 33.5, |
| "learning_rate": 0.000198, |
| "loss": 0.3074, |
| "step": 1072 |
| }, |
| { |
| "epoch": 33.53, |
| "learning_rate": 0.00019762499999999997, |
| "loss": 0.284, |
| "step": 1073 |
| }, |
| { |
| "epoch": 33.56, |
| "learning_rate": 0.00019724999999999997, |
| "loss": 0.2594, |
| "step": 1074 |
| }, |
| { |
| "epoch": 33.59, |
| "learning_rate": 0.00019687499999999997, |
| "loss": 0.289, |
| "step": 1075 |
| }, |
| { |
| "epoch": 33.62, |
| "learning_rate": 0.00019649999999999998, |
| "loss": 0.2332, |
| "step": 1076 |
| }, |
| { |
| "epoch": 33.66, |
| "learning_rate": 0.000196125, |
| "loss": 0.273, |
| "step": 1077 |
| }, |
| { |
| "epoch": 33.69, |
| "learning_rate": 0.00019574999999999996, |
| "loss": 0.2432, |
| "step": 1078 |
| }, |
| { |
| "epoch": 33.72, |
| "learning_rate": 0.000195375, |
| "loss": 0.2972, |
| "step": 1079 |
| }, |
| { |
| "epoch": 33.75, |
| "learning_rate": 0.000195, |
| "loss": 0.3204, |
| "step": 1080 |
| }, |
| { |
| "epoch": 33.75, |
| "eval_loss": 1.115964412689209, |
| "eval_mean_accuracy": 0.5335427145661599, |
| "eval_mean_iou": 0.40787664020271786, |
| "eval_overall_accuracy": 0.7002759900009423, |
| "eval_per_category_accuracy": [ |
| 0.6630361761260739, |
| 0.8082622146284658, |
| 0.7871573738335865, |
| 0.36711750785907105, |
| 0.5654822561268378, |
| 0.5633086423337531, |
| 0.5911038197336373, |
| 0.45939670699263163, |
| 0.372472338801984, |
| 0.5790521451098671, |
| 0.19228803261578203, |
| 0.4538353606322284 |
| ], |
| "eval_per_category_iou": [ |
| 0.5828634097547293, |
| 0.588827472393498, |
| 0.5929283087535941, |
| 0.28778222943917225, |
| 0.3846395528619974, |
| 0.4096045495929381, |
| 0.46128705870355435, |
| 0.4113942268943363, |
| 0.23692279038791758, |
| 0.4310282309880108, |
| 0.15848785287217135, |
| 0.3487539997906948 |
| ], |
| "eval_runtime": 123.6396, |
| "eval_samples_per_second": 0.922, |
| "eval_steps_per_second": 0.081, |
| "step": 1080 |
| }, |
| { |
| "epoch": 33.78, |
| "learning_rate": 0.000194625, |
| "loss": 0.2524, |
| "step": 1081 |
| }, |
| { |
| "epoch": 33.81, |
| "learning_rate": 0.00019424999999999998, |
| "loss": 0.2775, |
| "step": 1082 |
| }, |
| { |
| "epoch": 33.84, |
| "learning_rate": 0.00019387499999999998, |
| "loss": 0.2612, |
| "step": 1083 |
| }, |
| { |
| "epoch": 33.88, |
| "learning_rate": 0.0001935, |
| "loss": 0.2929, |
| "step": 1084 |
| }, |
| { |
| "epoch": 33.91, |
| "learning_rate": 0.000193125, |
| "loss": 0.3066, |
| "step": 1085 |
| }, |
| { |
| "epoch": 33.94, |
| "learning_rate": 0.00019274999999999997, |
| "loss": 0.2776, |
| "step": 1086 |
| }, |
| { |
| "epoch": 33.97, |
| "learning_rate": 0.00019237499999999997, |
| "loss": 0.3126, |
| "step": 1087 |
| }, |
| { |
| "epoch": 34.0, |
| "learning_rate": 0.00019199999999999998, |
| "loss": 0.2936, |
| "step": 1088 |
| }, |
| { |
| "epoch": 34.03, |
| "learning_rate": 0.00019162499999999998, |
| "loss": 0.2292, |
| "step": 1089 |
| }, |
| { |
| "epoch": 34.06, |
| "learning_rate": 0.00019124999999999996, |
| "loss": 0.2325, |
| "step": 1090 |
| }, |
| { |
| "epoch": 34.09, |
| "learning_rate": 0.000190875, |
| "loss": 0.3188, |
| "step": 1091 |
| }, |
| { |
| "epoch": 34.12, |
| "learning_rate": 0.0001905, |
| "loss": 0.2297, |
| "step": 1092 |
| }, |
| { |
| "epoch": 34.16, |
| "learning_rate": 0.000190125, |
| "loss": 0.2976, |
| "step": 1093 |
| }, |
| { |
| "epoch": 34.19, |
| "learning_rate": 0.00018974999999999998, |
| "loss": 0.2697, |
| "step": 1094 |
| }, |
| { |
| "epoch": 34.22, |
| "learning_rate": 0.00018937499999999998, |
| "loss": 0.2557, |
| "step": 1095 |
| }, |
| { |
| "epoch": 34.25, |
| "learning_rate": 0.00018899999999999999, |
| "loss": 0.2822, |
| "step": 1096 |
| }, |
| { |
| "epoch": 34.28, |
| "learning_rate": 0.000188625, |
| "loss": 0.2617, |
| "step": 1097 |
| }, |
| { |
| "epoch": 34.31, |
| "learning_rate": 0.00018824999999999997, |
| "loss": 0.3027, |
| "step": 1098 |
| }, |
| { |
| "epoch": 34.34, |
| "learning_rate": 0.00018787499999999997, |
| "loss": 0.2875, |
| "step": 1099 |
| }, |
| { |
| "epoch": 34.38, |
| "learning_rate": 0.00018749999999999998, |
| "loss": 0.2664, |
| "step": 1100 |
| }, |
| { |
| "epoch": 34.38, |
| "eval_loss": 1.1181931495666504, |
| "eval_mean_accuracy": 0.5435697888948486, |
| "eval_mean_iou": 0.4127189740214789, |
| "eval_overall_accuracy": 0.7017817246286493, |
| "eval_per_category_accuracy": [ |
| 0.724056287581135, |
| 0.7735105656570589, |
| 0.762912123737922, |
| 0.4073150082233989, |
| 0.5763937707528459, |
| 0.521946076598957, |
| 0.5840857814465582, |
| 0.48116329926466095, |
| 0.3559022386271883, |
| 0.6600358513602811, |
| 0.25081952661816337, |
| 0.4246969368700156 |
| ], |
| "eval_per_category_iou": [ |
| 0.6065393493247785, |
| 0.5813762518889343, |
| 0.6013614411734738, |
| 0.296553692868908, |
| 0.39027709190672155, |
| 0.3961876138082264, |
| 0.4634922320575209, |
| 0.4263561843601833, |
| 0.23952913777273507, |
| 0.4359275896836001, |
| 0.18188799857945087, |
| 0.33313910483321435 |
| ], |
| "eval_runtime": 123.7397, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 1100 |
| }, |
| { |
| "epoch": 34.41, |
| "learning_rate": 0.00018712499999999998, |
| "loss": 0.229, |
| "step": 1101 |
| }, |
| { |
| "epoch": 34.44, |
| "learning_rate": 0.00018675, |
| "loss": 0.2919, |
| "step": 1102 |
| }, |
| { |
| "epoch": 34.47, |
| "learning_rate": 0.00018637499999999996, |
| "loss": 0.268, |
| "step": 1103 |
| }, |
| { |
| "epoch": 34.5, |
| "learning_rate": 0.000186, |
| "loss": 0.2734, |
| "step": 1104 |
| }, |
| { |
| "epoch": 34.53, |
| "learning_rate": 0.000185625, |
| "loss": 0.2609, |
| "step": 1105 |
| }, |
| { |
| "epoch": 34.56, |
| "learning_rate": 0.00018525, |
| "loss": 0.2704, |
| "step": 1106 |
| }, |
| { |
| "epoch": 34.59, |
| "learning_rate": 0.00018487499999999998, |
| "loss": 0.2357, |
| "step": 1107 |
| }, |
| { |
| "epoch": 34.62, |
| "learning_rate": 0.00018449999999999999, |
| "loss": 0.2885, |
| "step": 1108 |
| }, |
| { |
| "epoch": 34.66, |
| "learning_rate": 0.000184125, |
| "loss": 0.279, |
| "step": 1109 |
| }, |
| { |
| "epoch": 34.69, |
| "learning_rate": 0.00018375, |
| "loss": 0.2706, |
| "step": 1110 |
| }, |
| { |
| "epoch": 34.72, |
| "learning_rate": 0.00018337499999999997, |
| "loss": 0.29, |
| "step": 1111 |
| }, |
| { |
| "epoch": 34.75, |
| "learning_rate": 0.00018299999999999998, |
| "loss": 0.2434, |
| "step": 1112 |
| }, |
| { |
| "epoch": 34.78, |
| "learning_rate": 0.00018262499999999998, |
| "loss": 0.2229, |
| "step": 1113 |
| }, |
| { |
| "epoch": 34.81, |
| "learning_rate": 0.00018224999999999998, |
| "loss": 0.3167, |
| "step": 1114 |
| }, |
| { |
| "epoch": 34.84, |
| "learning_rate": 0.00018187499999999996, |
| "loss": 0.2927, |
| "step": 1115 |
| }, |
| { |
| "epoch": 34.88, |
| "learning_rate": 0.00018149999999999997, |
| "loss": 0.2918, |
| "step": 1116 |
| }, |
| { |
| "epoch": 34.91, |
| "learning_rate": 0.000181125, |
| "loss": 0.2716, |
| "step": 1117 |
| }, |
| { |
| "epoch": 34.94, |
| "learning_rate": 0.00018075, |
| "loss": 0.2987, |
| "step": 1118 |
| }, |
| { |
| "epoch": 34.97, |
| "learning_rate": 0.00018037499999999998, |
| "loss": 0.2867, |
| "step": 1119 |
| }, |
| { |
| "epoch": 35.0, |
| "learning_rate": 0.00017999999999999998, |
| "loss": 0.2174, |
| "step": 1120 |
| }, |
| { |
| "epoch": 35.0, |
| "eval_loss": 1.146211862564087, |
| "eval_mean_accuracy": 0.5259360250815436, |
| "eval_mean_iou": 0.40902031856239435, |
| "eval_overall_accuracy": 0.7020502592387953, |
| "eval_per_category_accuracy": [ |
| 0.6636932884529285, |
| 0.8242473700456278, |
| 0.7637193856135033, |
| 0.33575061935725636, |
| 0.5172657600074553, |
| 0.5449210456595981, |
| 0.5798332359808486, |
| 0.5085132394239433, |
| 0.4039910206471877, |
| 0.5620809018228701, |
| 0.19158405083360625, |
| 0.41563238313369855 |
| ], |
| "eval_per_category_iou": [ |
| 0.581206368376567, |
| 0.590388972289196, |
| 0.6012892047747127, |
| 0.26994515396503915, |
| 0.39131440972408477, |
| 0.40628420438817353, |
| 0.46806224299219834, |
| 0.43803756630311286, |
| 0.2528141124145772, |
| 0.42486226723408954, |
| 0.15463729053074193, |
| 0.32940202975623956 |
| ], |
| "eval_runtime": 123.9909, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 1120 |
| }, |
| { |
| "epoch": 35.03, |
| "learning_rate": 0.000179625, |
| "loss": 0.3013, |
| "step": 1121 |
| }, |
| { |
| "epoch": 35.06, |
| "learning_rate": 0.00017925, |
| "loss": 0.3259, |
| "step": 1122 |
| }, |
| { |
| "epoch": 35.09, |
| "learning_rate": 0.00017887499999999997, |
| "loss": 0.2369, |
| "step": 1123 |
| }, |
| { |
| "epoch": 35.12, |
| "learning_rate": 0.00017849999999999997, |
| "loss": 0.2542, |
| "step": 1124 |
| }, |
| { |
| "epoch": 35.16, |
| "learning_rate": 0.00017812499999999998, |
| "loss": 0.276, |
| "step": 1125 |
| }, |
| { |
| "epoch": 35.19, |
| "learning_rate": 0.00017774999999999998, |
| "loss": 0.3107, |
| "step": 1126 |
| }, |
| { |
| "epoch": 35.22, |
| "learning_rate": 0.000177375, |
| "loss": 0.2636, |
| "step": 1127 |
| }, |
| { |
| "epoch": 35.25, |
| "learning_rate": 0.00017699999999999997, |
| "loss": 0.275, |
| "step": 1128 |
| }, |
| { |
| "epoch": 35.28, |
| "learning_rate": 0.00017662499999999997, |
| "loss": 0.2666, |
| "step": 1129 |
| }, |
| { |
| "epoch": 35.31, |
| "learning_rate": 0.00017625, |
| "loss": 0.2637, |
| "step": 1130 |
| }, |
| { |
| "epoch": 35.34, |
| "learning_rate": 0.000175875, |
| "loss": 0.3473, |
| "step": 1131 |
| }, |
| { |
| "epoch": 35.38, |
| "learning_rate": 0.00017549999999999998, |
| "loss": 0.282, |
| "step": 1132 |
| }, |
| { |
| "epoch": 35.41, |
| "learning_rate": 0.000175125, |
| "loss": 0.2527, |
| "step": 1133 |
| }, |
| { |
| "epoch": 35.44, |
| "learning_rate": 0.00017475, |
| "loss": 0.2537, |
| "step": 1134 |
| }, |
| { |
| "epoch": 35.47, |
| "learning_rate": 0.000174375, |
| "loss": 0.2478, |
| "step": 1135 |
| }, |
| { |
| "epoch": 35.5, |
| "learning_rate": 0.00017399999999999997, |
| "loss": 0.2486, |
| "step": 1136 |
| }, |
| { |
| "epoch": 35.53, |
| "learning_rate": 0.00017362499999999998, |
| "loss": 0.2905, |
| "step": 1137 |
| }, |
| { |
| "epoch": 35.56, |
| "learning_rate": 0.00017324999999999998, |
| "loss": 0.2614, |
| "step": 1138 |
| }, |
| { |
| "epoch": 35.59, |
| "learning_rate": 0.000172875, |
| "loss": 0.2829, |
| "step": 1139 |
| }, |
| { |
| "epoch": 35.62, |
| "learning_rate": 0.00017249999999999996, |
| "loss": 0.2166, |
| "step": 1140 |
| }, |
| { |
| "epoch": 35.62, |
| "eval_loss": 1.08646821975708, |
| "eval_mean_accuracy": 0.5490911489700377, |
| "eval_mean_iou": 0.4156727013035872, |
| "eval_overall_accuracy": 0.704173673663223, |
| "eval_per_category_accuracy": [ |
| 0.715107372154208, |
| 0.7864662671698294, |
| 0.7517882320108867, |
| 0.3890431017133816, |
| 0.5869568300287475, |
| 0.6103763895492108, |
| 0.6197900020310195, |
| 0.508113587003689, |
| 0.3439349440565025, |
| 0.5934411933902901, |
| 0.2247477143548007, |
| 0.45932815417788647 |
| ], |
| "eval_per_category_iou": [ |
| 0.603070687887534, |
| 0.5858622486005342, |
| 0.6031732004424101, |
| 0.289344450551861, |
| 0.392381679844199, |
| 0.41743040137801113, |
| 0.46797730602686066, |
| 0.4396605672232096, |
| 0.2309088769980208, |
| 0.43498577114869114, |
| 0.17453394151301732, |
| 0.34874328402869764 |
| ], |
| "eval_runtime": 123.1449, |
| "eval_samples_per_second": 0.926, |
| "eval_steps_per_second": 0.081, |
| "step": 1140 |
| }, |
| { |
| "epoch": 35.66, |
| "learning_rate": 0.00017212499999999997, |
| "loss": 0.2768, |
| "step": 1141 |
| }, |
| { |
| "epoch": 35.69, |
| "learning_rate": 0.00017175, |
| "loss": 0.235, |
| "step": 1142 |
| }, |
| { |
| "epoch": 35.72, |
| "learning_rate": 0.000171375, |
| "loss": 0.326, |
| "step": 1143 |
| }, |
| { |
| "epoch": 35.75, |
| "learning_rate": 0.00017099999999999998, |
| "loss": 0.2218, |
| "step": 1144 |
| }, |
| { |
| "epoch": 35.78, |
| "learning_rate": 0.000170625, |
| "loss": 0.2663, |
| "step": 1145 |
| }, |
| { |
| "epoch": 35.81, |
| "learning_rate": 0.00017025, |
| "loss": 0.2472, |
| "step": 1146 |
| }, |
| { |
| "epoch": 35.84, |
| "learning_rate": 0.000169875, |
| "loss": 0.295, |
| "step": 1147 |
| }, |
| { |
| "epoch": 35.88, |
| "learning_rate": 0.00016949999999999997, |
| "loss": 0.2811, |
| "step": 1148 |
| }, |
| { |
| "epoch": 35.91, |
| "learning_rate": 0.00016912499999999998, |
| "loss": 0.2804, |
| "step": 1149 |
| }, |
| { |
| "epoch": 35.94, |
| "learning_rate": 0.00016874999999999998, |
| "loss": 0.2542, |
| "step": 1150 |
| }, |
| { |
| "epoch": 35.97, |
| "learning_rate": 0.00016837499999999999, |
| "loss": 0.3409, |
| "step": 1151 |
| }, |
| { |
| "epoch": 36.0, |
| "learning_rate": 0.000168, |
| "loss": 0.3221, |
| "step": 1152 |
| }, |
| { |
| "epoch": 36.03, |
| "learning_rate": 0.00016762499999999997, |
| "loss": 0.2591, |
| "step": 1153 |
| }, |
| { |
| "epoch": 36.06, |
| "learning_rate": 0.00016724999999999997, |
| "loss": 0.2735, |
| "step": 1154 |
| }, |
| { |
| "epoch": 36.09, |
| "learning_rate": 0.000166875, |
| "loss": 0.2796, |
| "step": 1155 |
| }, |
| { |
| "epoch": 36.12, |
| "learning_rate": 0.0001665, |
| "loss": 0.2646, |
| "step": 1156 |
| }, |
| { |
| "epoch": 36.16, |
| "learning_rate": 0.00016612499999999999, |
| "loss": 0.2733, |
| "step": 1157 |
| }, |
| { |
| "epoch": 36.19, |
| "learning_rate": 0.00016575, |
| "loss": 0.2484, |
| "step": 1158 |
| }, |
| { |
| "epoch": 36.22, |
| "learning_rate": 0.000165375, |
| "loss": 0.2358, |
| "step": 1159 |
| }, |
| { |
| "epoch": 36.25, |
| "learning_rate": 0.000165, |
| "loss": 0.2663, |
| "step": 1160 |
| }, |
| { |
| "epoch": 36.25, |
| "eval_loss": 1.137108564376831, |
| "eval_mean_accuracy": 0.5305040371892099, |
| "eval_mean_iou": 0.40660775585344017, |
| "eval_overall_accuracy": 0.7020454741360849, |
| "eval_per_category_accuracy": [ |
| 0.6970544774752907, |
| 0.8085416177212705, |
| 0.746343708650657, |
| 0.36462272989084044, |
| 0.5652209146142579, |
| 0.5848210235211713, |
| 0.6542464838959048, |
| 0.4999008904855938, |
| 0.3446048463661127, |
| 0.5315337070066618, |
| 0.176665758632194, |
| 0.3924922880105646 |
| ], |
| "eval_per_category_iou": [ |
| 0.5935757656542305, |
| 0.5890707109807802, |
| 0.5950727233346256, |
| 0.2771828296203293, |
| 0.38909924647819555, |
| 0.412788314735688, |
| 0.4693415738953993, |
| 0.43332320989236645, |
| 0.23622363331143711, |
| 0.4156305897380057, |
| 0.15248277538253444, |
| 0.3155016972176909 |
| ], |
| "eval_runtime": 124.8642, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 1160 |
| }, |
| { |
| "epoch": 36.28, |
| "learning_rate": 0.00016462499999999998, |
| "loss": 0.2697, |
| "step": 1161 |
| }, |
| { |
| "epoch": 36.31, |
| "learning_rate": 0.00016424999999999998, |
| "loss": 0.3121, |
| "step": 1162 |
| }, |
| { |
| "epoch": 36.34, |
| "learning_rate": 0.00016387499999999999, |
| "loss": 0.2234, |
| "step": 1163 |
| }, |
| { |
| "epoch": 36.38, |
| "learning_rate": 0.0001635, |
| "loss": 0.2832, |
| "step": 1164 |
| }, |
| { |
| "epoch": 36.41, |
| "learning_rate": 0.00016312499999999997, |
| "loss": 0.2693, |
| "step": 1165 |
| }, |
| { |
| "epoch": 36.44, |
| "learning_rate": 0.00016274999999999997, |
| "loss": 0.2714, |
| "step": 1166 |
| }, |
| { |
| "epoch": 36.47, |
| "learning_rate": 0.00016237499999999998, |
| "loss": 0.2497, |
| "step": 1167 |
| }, |
| { |
| "epoch": 36.5, |
| "learning_rate": 0.000162, |
| "loss": 0.3223, |
| "step": 1168 |
| }, |
| { |
| "epoch": 36.53, |
| "learning_rate": 0.00016162499999999996, |
| "loss": 0.2469, |
| "step": 1169 |
| }, |
| { |
| "epoch": 36.56, |
| "learning_rate": 0.00016125, |
| "loss": 0.261, |
| "step": 1170 |
| }, |
| { |
| "epoch": 36.59, |
| "learning_rate": 0.000160875, |
| "loss": 0.2705, |
| "step": 1171 |
| }, |
| { |
| "epoch": 36.62, |
| "learning_rate": 0.0001605, |
| "loss": 0.2836, |
| "step": 1172 |
| }, |
| { |
| "epoch": 36.66, |
| "learning_rate": 0.00016012499999999998, |
| "loss": 0.293, |
| "step": 1173 |
| }, |
| { |
| "epoch": 36.69, |
| "learning_rate": 0.00015974999999999998, |
| "loss": 0.276, |
| "step": 1174 |
| }, |
| { |
| "epoch": 36.72, |
| "learning_rate": 0.00015937499999999998, |
| "loss": 0.3089, |
| "step": 1175 |
| }, |
| { |
| "epoch": 36.75, |
| "learning_rate": 0.000159, |
| "loss": 0.2605, |
| "step": 1176 |
| }, |
| { |
| "epoch": 36.78, |
| "learning_rate": 0.000158625, |
| "loss": 0.2842, |
| "step": 1177 |
| }, |
| { |
| "epoch": 36.81, |
| "learning_rate": 0.00015824999999999997, |
| "loss": 0.255, |
| "step": 1178 |
| }, |
| { |
| "epoch": 36.84, |
| "learning_rate": 0.00015787499999999997, |
| "loss": 0.2833, |
| "step": 1179 |
| }, |
| { |
| "epoch": 36.88, |
| "learning_rate": 0.00015749999999999998, |
| "loss": 0.238, |
| "step": 1180 |
| }, |
| { |
| "epoch": 36.88, |
| "eval_loss": 1.1475542783737183, |
| "eval_mean_accuracy": 0.5371533106205065, |
| "eval_mean_iou": 0.40893928799371637, |
| "eval_overall_accuracy": 0.7031427015338028, |
| "eval_per_category_accuracy": [ |
| 0.6976135001855333, |
| 0.7965799961937796, |
| 0.7460052807310956, |
| 0.3398449698475375, |
| 0.597345661629596, |
| 0.5292463135890019, |
| 0.6171465278226054, |
| 0.5018091504332425, |
| 0.37176250854014536, |
| 0.6093834181460872, |
| 0.1516437974613805, |
| 0.48745860286607445 |
| ], |
| "eval_per_category_iou": [ |
| 0.6004376500914973, |
| 0.5880653578769844, |
| 0.6009500892806243, |
| 0.27386734445646793, |
| 0.38869305463643816, |
| 0.3986490088311958, |
| 0.4687093632461693, |
| 0.4400306667769067, |
| 0.23646360665729813, |
| 0.43085623424494474, |
| 0.13374979415425067, |
| 0.3467992856718186 |
| ], |
| "eval_runtime": 125.7829, |
| "eval_samples_per_second": 0.906, |
| "eval_steps_per_second": 0.08, |
| "step": 1180 |
| }, |
| { |
| "epoch": 36.91, |
| "learning_rate": 0.000157125, |
| "loss": 0.2113, |
| "step": 1181 |
| }, |
| { |
| "epoch": 36.94, |
| "learning_rate": 0.00015675, |
| "loss": 0.2319, |
| "step": 1182 |
| }, |
| { |
| "epoch": 36.97, |
| "learning_rate": 0.000156375, |
| "loss": 0.2776, |
| "step": 1183 |
| }, |
| { |
| "epoch": 37.0, |
| "learning_rate": 0.000156, |
| "loss": 0.2724, |
| "step": 1184 |
| }, |
| { |
| "epoch": 37.03, |
| "learning_rate": 0.000155625, |
| "loss": 0.2522, |
| "step": 1185 |
| }, |
| { |
| "epoch": 37.06, |
| "learning_rate": 0.00015524999999999998, |
| "loss": 0.2685, |
| "step": 1186 |
| }, |
| { |
| "epoch": 37.09, |
| "learning_rate": 0.00015487499999999998, |
| "loss": 0.2625, |
| "step": 1187 |
| }, |
| { |
| "epoch": 37.12, |
| "learning_rate": 0.0001545, |
| "loss": 0.2931, |
| "step": 1188 |
| }, |
| { |
| "epoch": 37.16, |
| "learning_rate": 0.000154125, |
| "loss": 0.2407, |
| "step": 1189 |
| }, |
| { |
| "epoch": 37.19, |
| "learning_rate": 0.00015374999999999997, |
| "loss": 0.2495, |
| "step": 1190 |
| }, |
| { |
| "epoch": 37.22, |
| "learning_rate": 0.00015337499999999997, |
| "loss": 0.2707, |
| "step": 1191 |
| }, |
| { |
| "epoch": 37.25, |
| "learning_rate": 0.00015299999999999998, |
| "loss": 0.2649, |
| "step": 1192 |
| }, |
| { |
| "epoch": 37.28, |
| "learning_rate": 0.000152625, |
| "loss": 0.2394, |
| "step": 1193 |
| }, |
| { |
| "epoch": 37.31, |
| "learning_rate": 0.00015224999999999996, |
| "loss": 0.2631, |
| "step": 1194 |
| }, |
| { |
| "epoch": 37.34, |
| "learning_rate": 0.000151875, |
| "loss": 0.2567, |
| "step": 1195 |
| }, |
| { |
| "epoch": 37.38, |
| "learning_rate": 0.0001515, |
| "loss": 0.2468, |
| "step": 1196 |
| }, |
| { |
| "epoch": 37.41, |
| "learning_rate": 0.000151125, |
| "loss": 0.2228, |
| "step": 1197 |
| }, |
| { |
| "epoch": 37.44, |
| "learning_rate": 0.00015074999999999998, |
| "loss": 0.3007, |
| "step": 1198 |
| }, |
| { |
| "epoch": 37.47, |
| "learning_rate": 0.00015037499999999998, |
| "loss": 0.243, |
| "step": 1199 |
| }, |
| { |
| "epoch": 37.5, |
| "learning_rate": 0.00015, |
| "loss": 0.2841, |
| "step": 1200 |
| }, |
| { |
| "epoch": 37.5, |
| "eval_loss": 1.1354475021362305, |
| "eval_mean_accuracy": 0.5263534317011764, |
| "eval_mean_iou": 0.41013123193154694, |
| "eval_overall_accuracy": 0.7067963784201103, |
| "eval_per_category_accuracy": [ |
| 0.7139534975923257, |
| 0.804556808970909, |
| 0.7688788419487399, |
| 0.3723950562452724, |
| 0.5098590579144947, |
| 0.5213055139199896, |
| 0.6283580703110608, |
| 0.48324942061113607, |
| 0.3598040868477325, |
| 0.5770295734606055, |
| 0.17015239675188928, |
| 0.406698855839962 |
| ], |
| "eval_per_category_iou": [ |
| 0.607371665547802, |
| 0.5910491765871773, |
| 0.6018541516826075, |
| 0.28451868425517474, |
| 0.39342740031828116, |
| 0.40211358785036916, |
| 0.46925262506268345, |
| 0.4308380968392737, |
| 0.24084099244235252, |
| 0.42849446727985985, |
| 0.1475325433579703, |
| 0.32428139195501204 |
| ], |
| "eval_runtime": 124.1164, |
| "eval_samples_per_second": 0.918, |
| "eval_steps_per_second": 0.081, |
| "step": 1200 |
| }, |
| { |
| "epoch": 37.53, |
| "learning_rate": 0.000149625, |
| "loss": 0.2488, |
| "step": 1201 |
| }, |
| { |
| "epoch": 37.56, |
| "learning_rate": 0.00014925, |
| "loss": 0.2313, |
| "step": 1202 |
| }, |
| { |
| "epoch": 37.59, |
| "learning_rate": 0.000148875, |
| "loss": 0.2851, |
| "step": 1203 |
| }, |
| { |
| "epoch": 37.62, |
| "learning_rate": 0.00014849999999999998, |
| "loss": 0.2557, |
| "step": 1204 |
| }, |
| { |
| "epoch": 37.66, |
| "learning_rate": 0.00014812499999999998, |
| "loss": 0.2799, |
| "step": 1205 |
| }, |
| { |
| "epoch": 37.69, |
| "learning_rate": 0.00014774999999999999, |
| "loss": 0.2258, |
| "step": 1206 |
| }, |
| { |
| "epoch": 37.72, |
| "learning_rate": 0.000147375, |
| "loss": 0.2731, |
| "step": 1207 |
| }, |
| { |
| "epoch": 37.75, |
| "learning_rate": 0.000147, |
| "loss": 0.2759, |
| "step": 1208 |
| }, |
| { |
| "epoch": 37.78, |
| "learning_rate": 0.000146625, |
| "loss": 0.2671, |
| "step": 1209 |
| }, |
| { |
| "epoch": 37.81, |
| "learning_rate": 0.00014624999999999998, |
| "loss": 0.2322, |
| "step": 1210 |
| }, |
| { |
| "epoch": 37.84, |
| "learning_rate": 0.00014587499999999998, |
| "loss": 0.2694, |
| "step": 1211 |
| }, |
| { |
| "epoch": 37.88, |
| "learning_rate": 0.00014549999999999999, |
| "loss": 0.2515, |
| "step": 1212 |
| }, |
| { |
| "epoch": 37.91, |
| "learning_rate": 0.000145125, |
| "loss": 0.2551, |
| "step": 1213 |
| }, |
| { |
| "epoch": 37.94, |
| "learning_rate": 0.00014475, |
| "loss": 0.2726, |
| "step": 1214 |
| }, |
| { |
| "epoch": 37.97, |
| "learning_rate": 0.000144375, |
| "loss": 0.2333, |
| "step": 1215 |
| }, |
| { |
| "epoch": 38.0, |
| "learning_rate": 0.00014399999999999998, |
| "loss": 0.3171, |
| "step": 1216 |
| }, |
| { |
| "epoch": 38.03, |
| "learning_rate": 0.00014362499999999998, |
| "loss": 0.214, |
| "step": 1217 |
| }, |
| { |
| "epoch": 38.06, |
| "learning_rate": 0.00014324999999999999, |
| "loss": 0.2501, |
| "step": 1218 |
| }, |
| { |
| "epoch": 38.09, |
| "learning_rate": 0.000142875, |
| "loss": 0.27, |
| "step": 1219 |
| }, |
| { |
| "epoch": 38.12, |
| "learning_rate": 0.0001425, |
| "loss": 0.2818, |
| "step": 1220 |
| }, |
| { |
| "epoch": 38.12, |
| "eval_loss": 1.126877784729004, |
| "eval_mean_accuracy": 0.5428051261918013, |
| "eval_mean_iou": 0.4146794113167704, |
| "eval_overall_accuracy": 0.7030965905440482, |
| "eval_per_category_accuracy": [ |
| 0.6770392031011591, |
| 0.8094543668294201, |
| 0.758331362023993, |
| 0.37653798377526876, |
| 0.532966509794229, |
| 0.5858390282319381, |
| 0.5999678815522027, |
| 0.5306291257683666, |
| 0.36215983602920954, |
| 0.5728102323975506, |
| 0.26217505884369463, |
| 0.4457509259545844 |
| ], |
| "eval_per_category_iou": [ |
| 0.5905574420906589, |
| 0.5908683926326784, |
| 0.600062049082902, |
| 0.2877221170519082, |
| 0.38975146077536654, |
| 0.4145118899507094, |
| 0.46350329228443243, |
| 0.4431767133609311, |
| 0.24228546667101972, |
| 0.4268558960451029, |
| 0.1919931142844975, |
| 0.3348651015710377 |
| ], |
| "eval_runtime": 124.4527, |
| "eval_samples_per_second": 0.916, |
| "eval_steps_per_second": 0.08, |
| "step": 1220 |
| }, |
| { |
| "epoch": 38.16, |
| "learning_rate": 0.000142125, |
| "loss": 0.2966, |
| "step": 1221 |
| }, |
| { |
| "epoch": 38.19, |
| "learning_rate": 0.00014174999999999998, |
| "loss": 0.2926, |
| "step": 1222 |
| }, |
| { |
| "epoch": 38.22, |
| "learning_rate": 0.00014137499999999998, |
| "loss": 0.2629, |
| "step": 1223 |
| }, |
| { |
| "epoch": 38.25, |
| "learning_rate": 0.00014099999999999998, |
| "loss": 0.2444, |
| "step": 1224 |
| }, |
| { |
| "epoch": 38.28, |
| "learning_rate": 0.000140625, |
| "loss": 0.2438, |
| "step": 1225 |
| }, |
| { |
| "epoch": 38.31, |
| "learning_rate": 0.00014025, |
| "loss": 0.2376, |
| "step": 1226 |
| }, |
| { |
| "epoch": 38.34, |
| "learning_rate": 0.000139875, |
| "loss": 0.2408, |
| "step": 1227 |
| }, |
| { |
| "epoch": 38.38, |
| "learning_rate": 0.0001395, |
| "loss": 0.2635, |
| "step": 1228 |
| }, |
| { |
| "epoch": 38.41, |
| "learning_rate": 0.00013912499999999998, |
| "loss": 0.2601, |
| "step": 1229 |
| }, |
| { |
| "epoch": 38.44, |
| "learning_rate": 0.00013874999999999998, |
| "loss": 0.2365, |
| "step": 1230 |
| }, |
| { |
| "epoch": 38.47, |
| "learning_rate": 0.000138375, |
| "loss": 0.2792, |
| "step": 1231 |
| }, |
| { |
| "epoch": 38.5, |
| "learning_rate": 0.000138, |
| "loss": 0.2617, |
| "step": 1232 |
| }, |
| { |
| "epoch": 38.53, |
| "learning_rate": 0.000137625, |
| "loss": 0.2506, |
| "step": 1233 |
| }, |
| { |
| "epoch": 38.56, |
| "learning_rate": 0.00013725, |
| "loss": 0.2432, |
| "step": 1234 |
| }, |
| { |
| "epoch": 38.59, |
| "learning_rate": 0.00013687499999999998, |
| "loss": 0.2586, |
| "step": 1235 |
| }, |
| { |
| "epoch": 38.62, |
| "learning_rate": 0.00013649999999999998, |
| "loss": 0.2639, |
| "step": 1236 |
| }, |
| { |
| "epoch": 38.66, |
| "learning_rate": 0.000136125, |
| "loss": 0.2348, |
| "step": 1237 |
| }, |
| { |
| "epoch": 38.69, |
| "learning_rate": 0.00013575, |
| "loss": 0.2159, |
| "step": 1238 |
| }, |
| { |
| "epoch": 38.72, |
| "learning_rate": 0.00013537499999999997, |
| "loss": 0.262, |
| "step": 1239 |
| }, |
| { |
| "epoch": 38.75, |
| "learning_rate": 0.000135, |
| "loss": 0.2038, |
| "step": 1240 |
| }, |
| { |
| "epoch": 38.75, |
| "eval_loss": 1.1751402616500854, |
| "eval_mean_accuracy": 0.5304390484512038, |
| "eval_mean_iou": 0.40622936916347624, |
| "eval_overall_accuracy": 0.6986712740178693, |
| "eval_per_category_accuracy": [ |
| 0.6423987375220243, |
| 0.8134266444804227, |
| 0.7767237495074104, |
| 0.3901985413008931, |
| 0.5308068969848483, |
| 0.5436417349268162, |
| 0.6022397894037462, |
| 0.4768453387991784, |
| 0.38742313869196027, |
| 0.6249824628747533, |
| 0.14755764233440358, |
| 0.4290239045879889 |
| ], |
| "eval_per_category_iou": [ |
| 0.5715745466095057, |
| 0.5885654100627176, |
| 0.5987206381236377, |
| 0.29178201971401957, |
| 0.3945565323719773, |
| 0.40540716092251106, |
| 0.46018529546578313, |
| 0.4223820890139562, |
| 0.24166644758416572, |
| 0.43534810824439457, |
| 0.13081644614733368, |
| 0.3337477357017138 |
| ], |
| "eval_runtime": 124.2987, |
| "eval_samples_per_second": 0.917, |
| "eval_steps_per_second": 0.08, |
| "step": 1240 |
| }, |
| { |
| "epoch": 38.78, |
| "learning_rate": 0.00013462499999999998, |
| "loss": 0.2372, |
| "step": 1241 |
| }, |
| { |
| "epoch": 38.81, |
| "learning_rate": 0.00013424999999999998, |
| "loss": 0.248, |
| "step": 1242 |
| }, |
| { |
| "epoch": 38.84, |
| "learning_rate": 0.000133875, |
| "loss": 0.2933, |
| "step": 1243 |
| }, |
| { |
| "epoch": 38.88, |
| "learning_rate": 0.0001335, |
| "loss": 0.2774, |
| "step": 1244 |
| }, |
| { |
| "epoch": 38.91, |
| "learning_rate": 0.00013312499999999997, |
| "loss": 0.2443, |
| "step": 1245 |
| }, |
| { |
| "epoch": 38.94, |
| "learning_rate": 0.00013275, |
| "loss": 0.2406, |
| "step": 1246 |
| }, |
| { |
| "epoch": 38.97, |
| "learning_rate": 0.00013237499999999998, |
| "loss": 0.2329, |
| "step": 1247 |
| }, |
| { |
| "epoch": 39.0, |
| "learning_rate": 0.00013199999999999998, |
| "loss": 0.2936, |
| "step": 1248 |
| }, |
| { |
| "epoch": 39.03, |
| "learning_rate": 0.000131625, |
| "loss": 0.233, |
| "step": 1249 |
| }, |
| { |
| "epoch": 39.06, |
| "learning_rate": 0.00013125, |
| "loss": 0.2479, |
| "step": 1250 |
| }, |
| { |
| "epoch": 39.09, |
| "learning_rate": 0.000130875, |
| "loss": 0.2465, |
| "step": 1251 |
| }, |
| { |
| "epoch": 39.12, |
| "learning_rate": 0.0001305, |
| "loss": 0.2163, |
| "step": 1252 |
| }, |
| { |
| "epoch": 39.16, |
| "learning_rate": 0.000130125, |
| "loss": 0.2787, |
| "step": 1253 |
| }, |
| { |
| "epoch": 39.19, |
| "learning_rate": 0.00012974999999999998, |
| "loss": 0.2659, |
| "step": 1254 |
| }, |
| { |
| "epoch": 39.22, |
| "learning_rate": 0.000129375, |
| "loss": 0.2331, |
| "step": 1255 |
| }, |
| { |
| "epoch": 39.25, |
| "learning_rate": 0.000129, |
| "loss": 0.2257, |
| "step": 1256 |
| }, |
| { |
| "epoch": 39.28, |
| "learning_rate": 0.000128625, |
| "loss": 0.2665, |
| "step": 1257 |
| }, |
| { |
| "epoch": 39.31, |
| "learning_rate": 0.00012824999999999997, |
| "loss": 0.2628, |
| "step": 1258 |
| }, |
| { |
| "epoch": 39.34, |
| "learning_rate": 0.000127875, |
| "loss": 0.2324, |
| "step": 1259 |
| }, |
| { |
| "epoch": 39.38, |
| "learning_rate": 0.00012749999999999998, |
| "loss": 0.2283, |
| "step": 1260 |
| }, |
| { |
| "epoch": 39.38, |
| "eval_loss": 1.142168402671814, |
| "eval_mean_accuracy": 0.5394700525651125, |
| "eval_mean_iou": 0.4110475345037364, |
| "eval_overall_accuracy": 0.7037342807702851, |
| "eval_per_category_accuracy": [ |
| 0.6897202228631332, |
| 0.8044485564416076, |
| 0.7615044804603532, |
| 0.35452564520718105, |
| 0.5698946732927207, |
| 0.6162793651715365, |
| 0.5982407277662407, |
| 0.5245593376698656, |
| 0.3837431124282406, |
| 0.578711051283509, |
| 0.17813187721333404, |
| 0.4138815809836269 |
| ], |
| "eval_per_category_iou": [ |
| 0.5967777735850159, |
| 0.5893064258904264, |
| 0.6023111140803087, |
| 0.2766953088398972, |
| 0.38952115749696403, |
| 0.41883730710271455, |
| 0.4692414276858024, |
| 0.4456651655711127, |
| 0.23784259405654734, |
| 0.4280446932546577, |
| 0.151718492565017, |
| 0.3266089539163716 |
| ], |
| "eval_runtime": 124.9277, |
| "eval_samples_per_second": 0.913, |
| "eval_steps_per_second": 0.08, |
| "step": 1260 |
| }, |
| { |
| "epoch": 39.41, |
| "learning_rate": 0.000127125, |
| "loss": 0.2469, |
| "step": 1261 |
| }, |
| { |
| "epoch": 39.44, |
| "learning_rate": 0.00012675, |
| "loss": 0.2735, |
| "step": 1262 |
| }, |
| { |
| "epoch": 39.47, |
| "learning_rate": 0.000126375, |
| "loss": 0.2424, |
| "step": 1263 |
| }, |
| { |
| "epoch": 39.5, |
| "learning_rate": 0.00012599999999999997, |
| "loss": 0.3289, |
| "step": 1264 |
| }, |
| { |
| "epoch": 39.53, |
| "learning_rate": 0.000125625, |
| "loss": 0.2246, |
| "step": 1265 |
| }, |
| { |
| "epoch": 39.56, |
| "learning_rate": 0.00012524999999999998, |
| "loss": 0.2248, |
| "step": 1266 |
| }, |
| { |
| "epoch": 39.59, |
| "learning_rate": 0.00012487499999999999, |
| "loss": 0.2901, |
| "step": 1267 |
| }, |
| { |
| "epoch": 39.62, |
| "learning_rate": 0.0001245, |
| "loss": 0.2923, |
| "step": 1268 |
| }, |
| { |
| "epoch": 39.66, |
| "learning_rate": 0.000124125, |
| "loss": 0.2493, |
| "step": 1269 |
| }, |
| { |
| "epoch": 39.69, |
| "learning_rate": 0.00012374999999999997, |
| "loss": 0.2584, |
| "step": 1270 |
| }, |
| { |
| "epoch": 39.72, |
| "learning_rate": 0.000123375, |
| "loss": 0.2512, |
| "step": 1271 |
| }, |
| { |
| "epoch": 39.75, |
| "learning_rate": 0.00012299999999999998, |
| "loss": 0.2512, |
| "step": 1272 |
| }, |
| { |
| "epoch": 39.78, |
| "learning_rate": 0.00012262499999999999, |
| "loss": 0.2562, |
| "step": 1273 |
| }, |
| { |
| "epoch": 39.81, |
| "learning_rate": 0.00012225, |
| "loss": 0.2213, |
| "step": 1274 |
| }, |
| { |
| "epoch": 39.84, |
| "learning_rate": 0.000121875, |
| "loss": 0.2587, |
| "step": 1275 |
| }, |
| { |
| "epoch": 39.88, |
| "learning_rate": 0.0001215, |
| "loss": 0.2708, |
| "step": 1276 |
| }, |
| { |
| "epoch": 39.91, |
| "learning_rate": 0.00012112499999999999, |
| "loss": 0.26, |
| "step": 1277 |
| }, |
| { |
| "epoch": 39.94, |
| "learning_rate": 0.00012075, |
| "loss": 0.2967, |
| "step": 1278 |
| }, |
| { |
| "epoch": 39.97, |
| "learning_rate": 0.00012037499999999999, |
| "loss": 0.3146, |
| "step": 1279 |
| }, |
| { |
| "epoch": 40.0, |
| "learning_rate": 0.00011999999999999999, |
| "loss": 0.2447, |
| "step": 1280 |
| }, |
| { |
| "epoch": 40.0, |
| "eval_loss": 1.1354343891143799, |
| "eval_mean_accuracy": 0.5438332592868197, |
| "eval_mean_iou": 0.4150219791796939, |
| "eval_overall_accuracy": 0.7041060464423999, |
| "eval_per_category_accuracy": [ |
| 0.6808532445710408, |
| 0.799362142788886, |
| 0.7725039585508551, |
| 0.35654159235536187, |
| 0.5772243910641461, |
| 0.566482421726144, |
| 0.6123043569304211, |
| 0.5056481708669457, |
| 0.36814681064390475, |
| 0.6233672219263874, |
| 0.22334587237116368, |
| 0.4402189276465794 |
| ], |
| "eval_per_category_iou": [ |
| 0.5946733981307721, |
| 0.589769360751465, |
| 0.6016176962107231, |
| 0.2809759618934076, |
| 0.40105682771512385, |
| 0.41262175061825324, |
| 0.4680980262235374, |
| 0.44045844630382336, |
| 0.2373379933845958, |
| 0.4392363178265516, |
| 0.17845482625293776, |
| 0.3359631448451364 |
| ], |
| "eval_runtime": 123.6998, |
| "eval_samples_per_second": 0.922, |
| "eval_steps_per_second": 0.081, |
| "step": 1280 |
| }, |
| { |
| "epoch": 40.03, |
| "learning_rate": 0.000119625, |
| "loss": 0.2456, |
| "step": 1281 |
| }, |
| { |
| "epoch": 40.06, |
| "learning_rate": 0.00011925, |
| "loss": 0.2285, |
| "step": 1282 |
| }, |
| { |
| "epoch": 40.09, |
| "learning_rate": 0.00011887499999999999, |
| "loss": 0.2609, |
| "step": 1283 |
| }, |
| { |
| "epoch": 40.12, |
| "learning_rate": 0.0001185, |
| "loss": 0.2672, |
| "step": 1284 |
| }, |
| { |
| "epoch": 40.16, |
| "learning_rate": 0.00011812499999999998, |
| "loss": 0.2329, |
| "step": 1285 |
| }, |
| { |
| "epoch": 40.19, |
| "learning_rate": 0.00011774999999999999, |
| "loss": 0.2051, |
| "step": 1286 |
| }, |
| { |
| "epoch": 40.22, |
| "learning_rate": 0.00011737499999999998, |
| "loss": 0.2995, |
| "step": 1287 |
| }, |
| { |
| "epoch": 40.25, |
| "learning_rate": 0.000117, |
| "loss": 0.1885, |
| "step": 1288 |
| }, |
| { |
| "epoch": 40.28, |
| "learning_rate": 0.00011662499999999999, |
| "loss": 0.2992, |
| "step": 1289 |
| }, |
| { |
| "epoch": 40.31, |
| "learning_rate": 0.00011624999999999999, |
| "loss": 0.2432, |
| "step": 1290 |
| }, |
| { |
| "epoch": 40.34, |
| "learning_rate": 0.00011587499999999998, |
| "loss": 0.2271, |
| "step": 1291 |
| }, |
| { |
| "epoch": 40.38, |
| "learning_rate": 0.00011549999999999999, |
| "loss": 0.213, |
| "step": 1292 |
| }, |
| { |
| "epoch": 40.41, |
| "learning_rate": 0.00011512499999999998, |
| "loss": 0.2206, |
| "step": 1293 |
| }, |
| { |
| "epoch": 40.44, |
| "learning_rate": 0.00011475, |
| "loss": 0.2225, |
| "step": 1294 |
| }, |
| { |
| "epoch": 40.47, |
| "learning_rate": 0.00011437499999999999, |
| "loss": 0.2131, |
| "step": 1295 |
| }, |
| { |
| "epoch": 40.5, |
| "learning_rate": 0.00011399999999999999, |
| "loss": 0.2423, |
| "step": 1296 |
| }, |
| { |
| "epoch": 40.53, |
| "learning_rate": 0.00011362499999999998, |
| "loss": 0.271, |
| "step": 1297 |
| }, |
| { |
| "epoch": 40.56, |
| "learning_rate": 0.00011324999999999999, |
| "loss": 0.2445, |
| "step": 1298 |
| }, |
| { |
| "epoch": 40.59, |
| "learning_rate": 0.00011287499999999998, |
| "loss": 0.2475, |
| "step": 1299 |
| }, |
| { |
| "epoch": 40.62, |
| "learning_rate": 0.0001125, |
| "loss": 0.2321, |
| "step": 1300 |
| }, |
| { |
| "epoch": 40.62, |
| "eval_loss": 1.1653814315795898, |
| "eval_mean_accuracy": 0.5408112483993465, |
| "eval_mean_iou": 0.41269294807560053, |
| "eval_overall_accuracy": 0.7036235541628119, |
| "eval_per_category_accuracy": [ |
| 0.6838142881623307, |
| 0.8093913879494158, |
| 0.7532018676883804, |
| 0.37791548983004974, |
| 0.5957067059421767, |
| 0.573243715045783, |
| 0.5930246288666124, |
| 0.4893802815995954, |
| 0.3594425170581085, |
| 0.5970049458604821, |
| 0.23904466611368388, |
| 0.4185644866755393 |
| ], |
| "eval_per_category_iou": [ |
| 0.5959056561376641, |
| 0.59042517085335, |
| 0.6023650968501204, |
| 0.2890122487103827, |
| 0.3906925759840558, |
| 0.4110663193206453, |
| 0.46506967544061106, |
| 0.43331426440913373, |
| 0.23353692985288013, |
| 0.43202222849577215, |
| 0.18389822127407074, |
| 0.3250069895785184 |
| ], |
| "eval_runtime": 124.7159, |
| "eval_samples_per_second": 0.914, |
| "eval_steps_per_second": 0.08, |
| "step": 1300 |
| }, |
| { |
| "epoch": 40.66, |
| "learning_rate": 0.000112125, |
| "loss": 0.3208, |
| "step": 1301 |
| }, |
| { |
| "epoch": 40.69, |
| "learning_rate": 0.00011174999999999999, |
| "loss": 0.3019, |
| "step": 1302 |
| }, |
| { |
| "epoch": 40.72, |
| "learning_rate": 0.000111375, |
| "loss": 0.2665, |
| "step": 1303 |
| }, |
| { |
| "epoch": 40.75, |
| "learning_rate": 0.00011099999999999999, |
| "loss": 0.2561, |
| "step": 1304 |
| }, |
| { |
| "epoch": 40.78, |
| "learning_rate": 0.00011062499999999999, |
| "loss": 0.2655, |
| "step": 1305 |
| }, |
| { |
| "epoch": 40.81, |
| "learning_rate": 0.00011024999999999998, |
| "loss": 0.2366, |
| "step": 1306 |
| }, |
| { |
| "epoch": 40.84, |
| "learning_rate": 0.000109875, |
| "loss": 0.2378, |
| "step": 1307 |
| }, |
| { |
| "epoch": 40.88, |
| "learning_rate": 0.00010949999999999999, |
| "loss": 0.271, |
| "step": 1308 |
| }, |
| { |
| "epoch": 40.91, |
| "learning_rate": 0.000109125, |
| "loss": 0.2358, |
| "step": 1309 |
| }, |
| { |
| "epoch": 40.94, |
| "learning_rate": 0.00010874999999999999, |
| "loss": 0.2508, |
| "step": 1310 |
| }, |
| { |
| "epoch": 40.97, |
| "learning_rate": 0.00010837499999999999, |
| "loss": 0.2402, |
| "step": 1311 |
| }, |
| { |
| "epoch": 41.0, |
| "learning_rate": 0.00010799999999999998, |
| "loss": 0.2611, |
| "step": 1312 |
| }, |
| { |
| "epoch": 41.03, |
| "learning_rate": 0.000107625, |
| "loss": 0.2451, |
| "step": 1313 |
| }, |
| { |
| "epoch": 41.06, |
| "learning_rate": 0.00010724999999999999, |
| "loss": 0.2598, |
| "step": 1314 |
| }, |
| { |
| "epoch": 41.09, |
| "learning_rate": 0.000106875, |
| "loss": 0.2569, |
| "step": 1315 |
| }, |
| { |
| "epoch": 41.12, |
| "learning_rate": 0.00010649999999999999, |
| "loss": 0.2569, |
| "step": 1316 |
| }, |
| { |
| "epoch": 41.16, |
| "learning_rate": 0.00010612499999999999, |
| "loss": 0.2237, |
| "step": 1317 |
| }, |
| { |
| "epoch": 41.19, |
| "learning_rate": 0.00010574999999999998, |
| "loss": 0.202, |
| "step": 1318 |
| }, |
| { |
| "epoch": 41.22, |
| "learning_rate": 0.00010537499999999999, |
| "loss": 0.2057, |
| "step": 1319 |
| }, |
| { |
| "epoch": 41.25, |
| "learning_rate": 0.00010499999999999999, |
| "loss": 0.2372, |
| "step": 1320 |
| }, |
| { |
| "epoch": 41.25, |
| "eval_loss": 1.1355468034744263, |
| "eval_mean_accuracy": 0.5412489804633339, |
| "eval_mean_iou": 0.41447692698190547, |
| "eval_overall_accuracy": 0.7077278337980571, |
| "eval_per_category_accuracy": [ |
| 0.727894264092196, |
| 0.7965179874634031, |
| 0.7498001819977463, |
| 0.3499142962227882, |
| 0.5977569199788496, |
| 0.5723273293435776, |
| 0.6221028451591674, |
| 0.5215528371571748, |
| 0.368428524529072, |
| 0.574118280282117, |
| 0.2145154921903934, |
| 0.4000588071435204 |
| ], |
| "eval_per_category_iou": [ |
| 0.6142246674850493, |
| 0.5909511157539044, |
| 0.6008111638609991, |
| 0.27687830324661955, |
| 0.3943626392867024, |
| 0.41228150441043443, |
| 0.47144955393843074, |
| 0.4455374043062639, |
| 0.24181065357535947, |
| 0.4295235601706089, |
| 0.17377322327819833, |
| 0.3221193344702951 |
| ], |
| "eval_runtime": 124.519, |
| "eval_samples_per_second": 0.916, |
| "eval_steps_per_second": 0.08, |
| "step": 1320 |
| }, |
| { |
| "epoch": 41.28, |
| "learning_rate": 0.000104625, |
| "loss": 0.2561, |
| "step": 1321 |
| }, |
| { |
| "epoch": 41.31, |
| "learning_rate": 0.00010424999999999999, |
| "loss": 0.2621, |
| "step": 1322 |
| }, |
| { |
| "epoch": 41.34, |
| "learning_rate": 0.00010387499999999999, |
| "loss": 0.2444, |
| "step": 1323 |
| }, |
| { |
| "epoch": 41.38, |
| "learning_rate": 0.00010349999999999998, |
| "loss": 0.2312, |
| "step": 1324 |
| }, |
| { |
| "epoch": 41.41, |
| "learning_rate": 0.00010312499999999999, |
| "loss": 0.2854, |
| "step": 1325 |
| }, |
| { |
| "epoch": 41.44, |
| "learning_rate": 0.00010275, |
| "loss": 0.2313, |
| "step": 1326 |
| }, |
| { |
| "epoch": 41.47, |
| "learning_rate": 0.000102375, |
| "loss": 0.2526, |
| "step": 1327 |
| }, |
| { |
| "epoch": 41.5, |
| "learning_rate": 0.000102, |
| "loss": 0.2466, |
| "step": 1328 |
| }, |
| { |
| "epoch": 41.53, |
| "learning_rate": 0.00010162499999999999, |
| "loss": 0.2007, |
| "step": 1329 |
| }, |
| { |
| "epoch": 41.56, |
| "learning_rate": 0.00010125, |
| "loss": 0.252, |
| "step": 1330 |
| }, |
| { |
| "epoch": 41.59, |
| "learning_rate": 0.00010087499999999999, |
| "loss": 0.2439, |
| "step": 1331 |
| }, |
| { |
| "epoch": 41.62, |
| "learning_rate": 0.0001005, |
| "loss": 0.2105, |
| "step": 1332 |
| }, |
| { |
| "epoch": 41.66, |
| "learning_rate": 0.000100125, |
| "loss": 0.2287, |
| "step": 1333 |
| }, |
| { |
| "epoch": 41.69, |
| "learning_rate": 9.975e-05, |
| "loss": 0.2218, |
| "step": 1334 |
| }, |
| { |
| "epoch": 41.72, |
| "learning_rate": 9.937499999999999e-05, |
| "loss": 0.2428, |
| "step": 1335 |
| }, |
| { |
| "epoch": 41.75, |
| "learning_rate": 9.9e-05, |
| "loss": 0.2563, |
| "step": 1336 |
| }, |
| { |
| "epoch": 41.78, |
| "learning_rate": 9.862499999999998e-05, |
| "loss": 0.2548, |
| "step": 1337 |
| }, |
| { |
| "epoch": 41.81, |
| "learning_rate": 9.824999999999999e-05, |
| "loss": 0.2652, |
| "step": 1338 |
| }, |
| { |
| "epoch": 41.84, |
| "learning_rate": 9.787499999999998e-05, |
| "loss": 0.2626, |
| "step": 1339 |
| }, |
| { |
| "epoch": 41.88, |
| "learning_rate": 9.75e-05, |
| "loss": 0.2445, |
| "step": 1340 |
| }, |
| { |
| "epoch": 41.88, |
| "eval_loss": 1.1469475030899048, |
| "eval_mean_accuracy": 0.5379484863780519, |
| "eval_mean_iou": 0.41493570773872007, |
| "eval_overall_accuracy": 0.707922149122807, |
| "eval_per_category_accuracy": [ |
| 0.7132913926801945, |
| 0.8034154279338093, |
| 0.7410501366980564, |
| 0.35646872679578906, |
| 0.5641775744671368, |
| 0.5736120839518181, |
| 0.6049163267201869, |
| 0.523083945222975, |
| 0.33929886515886887, |
| 0.606417883747505, |
| 0.18094168276132264, |
| 0.44870779039896 |
| ], |
| "eval_per_category_iou": [ |
| 0.6081059511718557, |
| 0.5921678482858844, |
| 0.6024023582386882, |
| 0.2834983981036638, |
| 0.3992539132279866, |
| 0.41145093710666447, |
| 0.4672568610246129, |
| 0.4481674431413949, |
| 0.23338917828577357, |
| 0.43708494256390745, |
| 0.15454557337620747, |
| 0.3419050883380004 |
| ], |
| "eval_runtime": 125.6662, |
| "eval_samples_per_second": 0.907, |
| "eval_steps_per_second": 0.08, |
| "step": 1340 |
| }, |
| { |
| "epoch": 41.91, |
| "learning_rate": 9.712499999999999e-05, |
| "loss": 0.229, |
| "step": 1341 |
| }, |
| { |
| "epoch": 41.94, |
| "learning_rate": 9.675e-05, |
| "loss": 0.2578, |
| "step": 1342 |
| }, |
| { |
| "epoch": 41.97, |
| "learning_rate": 9.637499999999998e-05, |
| "loss": 0.2072, |
| "step": 1343 |
| }, |
| { |
| "epoch": 42.0, |
| "learning_rate": 9.599999999999999e-05, |
| "loss": 0.2624, |
| "step": 1344 |
| }, |
| { |
| "epoch": 42.03, |
| "learning_rate": 9.562499999999998e-05, |
| "loss": 0.2277, |
| "step": 1345 |
| }, |
| { |
| "epoch": 42.06, |
| "learning_rate": 9.525e-05, |
| "loss": 0.2581, |
| "step": 1346 |
| }, |
| { |
| "epoch": 42.09, |
| "learning_rate": 9.487499999999999e-05, |
| "loss": 0.2763, |
| "step": 1347 |
| }, |
| { |
| "epoch": 42.12, |
| "learning_rate": 9.449999999999999e-05, |
| "loss": 0.2368, |
| "step": 1348 |
| }, |
| { |
| "epoch": 42.16, |
| "learning_rate": 9.412499999999998e-05, |
| "loss": 0.2615, |
| "step": 1349 |
| }, |
| { |
| "epoch": 42.19, |
| "learning_rate": 9.374999999999999e-05, |
| "loss": 0.2606, |
| "step": 1350 |
| }, |
| { |
| "epoch": 42.22, |
| "learning_rate": 9.3375e-05, |
| "loss": 0.2877, |
| "step": 1351 |
| }, |
| { |
| "epoch": 42.25, |
| "learning_rate": 9.3e-05, |
| "loss": 0.234, |
| "step": 1352 |
| }, |
| { |
| "epoch": 42.28, |
| "learning_rate": 9.2625e-05, |
| "loss": 0.2459, |
| "step": 1353 |
| }, |
| { |
| "epoch": 42.31, |
| "learning_rate": 9.224999999999999e-05, |
| "loss": 0.2455, |
| "step": 1354 |
| }, |
| { |
| "epoch": 42.34, |
| "learning_rate": 9.1875e-05, |
| "loss": 0.21, |
| "step": 1355 |
| }, |
| { |
| "epoch": 42.38, |
| "learning_rate": 9.149999999999999e-05, |
| "loss": 0.2474, |
| "step": 1356 |
| }, |
| { |
| "epoch": 42.41, |
| "learning_rate": 9.112499999999999e-05, |
| "loss": 0.1975, |
| "step": 1357 |
| }, |
| { |
| "epoch": 42.44, |
| "learning_rate": 9.074999999999998e-05, |
| "loss": 0.2478, |
| "step": 1358 |
| }, |
| { |
| "epoch": 42.47, |
| "learning_rate": 9.0375e-05, |
| "loss": 0.2257, |
| "step": 1359 |
| }, |
| { |
| "epoch": 42.5, |
| "learning_rate": 8.999999999999999e-05, |
| "loss": 0.2302, |
| "step": 1360 |
| }, |
| { |
| "epoch": 42.5, |
| "eval_loss": 1.153663158416748, |
| "eval_mean_accuracy": 0.5419531448888287, |
| "eval_mean_iou": 0.41515638588403503, |
| "eval_overall_accuracy": 0.7042664645010965, |
| "eval_per_category_accuracy": [ |
| 0.6827689289098441, |
| 0.810421444316726, |
| 0.753038360776147, |
| 0.36378304106147774, |
| 0.5584057762551989, |
| 0.556503798010445, |
| 0.6197569389229929, |
| 0.5185966949923445, |
| 0.41072109881724533, |
| 0.5862469300251753, |
| 0.24506830153682282, |
| 0.3981264250415261 |
| ], |
| "eval_per_category_iou": [ |
| 0.5945439871028722, |
| 0.5908011746140732, |
| 0.6008308340237598, |
| 0.27827455601827156, |
| 0.4007411952172699, |
| 0.40577602711903876, |
| 0.47168119059245034, |
| 0.4496326388043784, |
| 0.2505815975629066, |
| 0.43150905536098794, |
| 0.18712751884900694, |
| 0.3203768553434056 |
| ], |
| "eval_runtime": 125.8859, |
| "eval_samples_per_second": 0.906, |
| "eval_steps_per_second": 0.079, |
| "step": 1360 |
| }, |
| { |
| "epoch": 42.53, |
| "learning_rate": 8.9625e-05, |
| "loss": 0.2433, |
| "step": 1361 |
| }, |
| { |
| "epoch": 42.56, |
| "learning_rate": 8.924999999999999e-05, |
| "loss": 0.219, |
| "step": 1362 |
| }, |
| { |
| "epoch": 42.59, |
| "learning_rate": 8.887499999999999e-05, |
| "loss": 0.2339, |
| "step": 1363 |
| }, |
| { |
| "epoch": 42.62, |
| "learning_rate": 8.849999999999998e-05, |
| "loss": 0.2808, |
| "step": 1364 |
| }, |
| { |
| "epoch": 42.66, |
| "learning_rate": 8.8125e-05, |
| "loss": 0.2129, |
| "step": 1365 |
| }, |
| { |
| "epoch": 42.69, |
| "learning_rate": 8.774999999999999e-05, |
| "loss": 0.2239, |
| "step": 1366 |
| }, |
| { |
| "epoch": 42.72, |
| "learning_rate": 8.7375e-05, |
| "loss": 0.2573, |
| "step": 1367 |
| }, |
| { |
| "epoch": 42.75, |
| "learning_rate": 8.699999999999999e-05, |
| "loss": 0.2332, |
| "step": 1368 |
| }, |
| { |
| "epoch": 42.78, |
| "learning_rate": 8.662499999999999e-05, |
| "loss": 0.2512, |
| "step": 1369 |
| }, |
| { |
| "epoch": 42.81, |
| "learning_rate": 8.624999999999998e-05, |
| "loss": 0.27, |
| "step": 1370 |
| }, |
| { |
| "epoch": 42.84, |
| "learning_rate": 8.5875e-05, |
| "loss": 0.2408, |
| "step": 1371 |
| }, |
| { |
| "epoch": 42.88, |
| "learning_rate": 8.549999999999999e-05, |
| "loss": 0.2451, |
| "step": 1372 |
| }, |
| { |
| "epoch": 42.91, |
| "learning_rate": 8.5125e-05, |
| "loss": 0.2501, |
| "step": 1373 |
| }, |
| { |
| "epoch": 42.94, |
| "learning_rate": 8.474999999999999e-05, |
| "loss": 0.2536, |
| "step": 1374 |
| }, |
| { |
| "epoch": 42.97, |
| "learning_rate": 8.437499999999999e-05, |
| "loss": 0.2217, |
| "step": 1375 |
| }, |
| { |
| "epoch": 43.0, |
| "learning_rate": 8.4e-05, |
| "loss": 0.2146, |
| "step": 1376 |
| }, |
| { |
| "epoch": 43.03, |
| "learning_rate": 8.362499999999999e-05, |
| "loss": 0.2469, |
| "step": 1377 |
| }, |
| { |
| "epoch": 43.06, |
| "learning_rate": 8.325e-05, |
| "loss": 0.2542, |
| "step": 1378 |
| }, |
| { |
| "epoch": 43.09, |
| "learning_rate": 8.2875e-05, |
| "loss": 0.2323, |
| "step": 1379 |
| }, |
| { |
| "epoch": 43.12, |
| "learning_rate": 8.25e-05, |
| "loss": 0.2529, |
| "step": 1380 |
| }, |
| { |
| "epoch": 43.12, |
| "eval_loss": 1.178342580795288, |
| "eval_mean_accuracy": 0.5386598224094629, |
| "eval_mean_iou": 0.41235210095330266, |
| "eval_overall_accuracy": 0.7036323881985849, |
| "eval_per_category_accuracy": [ |
| 0.6848815266855885, |
| 0.8030204961894948, |
| 0.7550909004265732, |
| 0.3582140304369852, |
| 0.5753484047025265, |
| 0.555687216691648, |
| 0.640783501194208, |
| 0.49069388446080675, |
| 0.36215983602920954, |
| 0.6092013971286576, |
| 0.18929764043671357, |
| 0.439539034531142 |
| ], |
| "eval_per_category_iou": [ |
| 0.594133983706034, |
| 0.5893338657431374, |
| 0.6019369021394456, |
| 0.2793258603289529, |
| 0.4035621970576491, |
| 0.4083979595238889, |
| 0.4713398441888125, |
| 0.433936019617467, |
| 0.2389563934512606, |
| 0.4345312098925259, |
| 0.1578312053877897, |
| 0.334939770402668 |
| ], |
| "eval_runtime": 126.51, |
| "eval_samples_per_second": 0.901, |
| "eval_steps_per_second": 0.079, |
| "step": 1380 |
| }, |
| { |
| "epoch": 43.16, |
| "learning_rate": 8.212499999999999e-05, |
| "loss": 0.2186, |
| "step": 1381 |
| }, |
| { |
| "epoch": 43.19, |
| "learning_rate": 8.175e-05, |
| "loss": 0.2749, |
| "step": 1382 |
| }, |
| { |
| "epoch": 43.22, |
| "learning_rate": 8.137499999999999e-05, |
| "loss": 0.2076, |
| "step": 1383 |
| }, |
| { |
| "epoch": 43.25, |
| "learning_rate": 8.1e-05, |
| "loss": 0.2138, |
| "step": 1384 |
| }, |
| { |
| "epoch": 43.28, |
| "learning_rate": 8.0625e-05, |
| "loss": 0.2677, |
| "step": 1385 |
| }, |
| { |
| "epoch": 43.31, |
| "learning_rate": 8.025e-05, |
| "loss": 0.2394, |
| "step": 1386 |
| }, |
| { |
| "epoch": 43.34, |
| "learning_rate": 7.987499999999999e-05, |
| "loss": 0.2616, |
| "step": 1387 |
| }, |
| { |
| "epoch": 43.38, |
| "learning_rate": 7.95e-05, |
| "loss": 0.2276, |
| "step": 1388 |
| }, |
| { |
| "epoch": 43.41, |
| "learning_rate": 7.912499999999999e-05, |
| "loss": 0.2384, |
| "step": 1389 |
| }, |
| { |
| "epoch": 43.44, |
| "learning_rate": 7.874999999999999e-05, |
| "loss": 0.2149, |
| "step": 1390 |
| }, |
| { |
| "epoch": 43.47, |
| "learning_rate": 7.8375e-05, |
| "loss": 0.245, |
| "step": 1391 |
| }, |
| { |
| "epoch": 43.5, |
| "learning_rate": 7.8e-05, |
| "loss": 0.2719, |
| "step": 1392 |
| }, |
| { |
| "epoch": 43.53, |
| "learning_rate": 7.762499999999999e-05, |
| "loss": 0.2792, |
| "step": 1393 |
| }, |
| { |
| "epoch": 43.56, |
| "learning_rate": 7.725e-05, |
| "loss": 0.222, |
| "step": 1394 |
| }, |
| { |
| "epoch": 43.59, |
| "learning_rate": 7.687499999999998e-05, |
| "loss": 0.245, |
| "step": 1395 |
| }, |
| { |
| "epoch": 43.62, |
| "learning_rate": 7.649999999999999e-05, |
| "loss": 0.2453, |
| "step": 1396 |
| }, |
| { |
| "epoch": 43.66, |
| "learning_rate": 7.612499999999998e-05, |
| "loss": 0.2785, |
| "step": 1397 |
| }, |
| { |
| "epoch": 43.69, |
| "learning_rate": 7.575e-05, |
| "loss": 0.2282, |
| "step": 1398 |
| }, |
| { |
| "epoch": 43.72, |
| "learning_rate": 7.537499999999999e-05, |
| "loss": 0.2474, |
| "step": 1399 |
| }, |
| { |
| "epoch": 43.75, |
| "learning_rate": 7.5e-05, |
| "loss": 0.2663, |
| "step": 1400 |
| }, |
| { |
| "epoch": 43.75, |
| "eval_loss": 1.1498355865478516, |
| "eval_mean_accuracy": 0.5427016091156777, |
| "eval_mean_iou": 0.41541271657914375, |
| "eval_overall_accuracy": 0.7057312078643263, |
| "eval_per_category_accuracy": [ |
| 0.6906194755759058, |
| 0.809933054824934, |
| 0.7446159000056786, |
| 0.37995225570953706, |
| 0.5716308723336582, |
| 0.5887787209796073, |
| 0.6002434074524245, |
| 0.5295512428333107, |
| 0.36392110236639663, |
| 0.6048871339304388, |
| 0.2275360943702883, |
| 0.40075004900595296 |
| ], |
| "eval_per_category_iou": [ |
| 0.5982699073297297, |
| 0.5925399284841079, |
| 0.600599711054169, |
| 0.2848265726808079, |
| 0.40146293851162157, |
| 0.4144918797593757, |
| 0.4681804226891479, |
| 0.4450808282475161, |
| 0.23986326945654288, |
| 0.43628356567819004, |
| 0.1795662718120157, |
| 0.32378730324650046 |
| ], |
| "eval_runtime": 124.214, |
| "eval_samples_per_second": 0.918, |
| "eval_steps_per_second": 0.081, |
| "step": 1400 |
| }, |
| { |
| "epoch": 43.78, |
| "learning_rate": 7.4625e-05, |
| "loss": 0.2215, |
| "step": 1401 |
| }, |
| { |
| "epoch": 43.81, |
| "learning_rate": 7.424999999999999e-05, |
| "loss": 0.2691, |
| "step": 1402 |
| }, |
| { |
| "epoch": 43.84, |
| "learning_rate": 7.387499999999999e-05, |
| "loss": 0.2965, |
| "step": 1403 |
| }, |
| { |
| "epoch": 43.88, |
| "learning_rate": 7.35e-05, |
| "loss": 0.2368, |
| "step": 1404 |
| }, |
| { |
| "epoch": 43.91, |
| "learning_rate": 7.312499999999999e-05, |
| "loss": 0.2511, |
| "step": 1405 |
| }, |
| { |
| "epoch": 43.94, |
| "learning_rate": 7.274999999999999e-05, |
| "loss": 0.2525, |
| "step": 1406 |
| }, |
| { |
| "epoch": 43.97, |
| "learning_rate": 7.2375e-05, |
| "loss": 0.235, |
| "step": 1407 |
| }, |
| { |
| "epoch": 44.0, |
| "learning_rate": 7.199999999999999e-05, |
| "loss": 0.2263, |
| "step": 1408 |
| }, |
| { |
| "epoch": 44.03, |
| "learning_rate": 7.162499999999999e-05, |
| "loss": 0.2373, |
| "step": 1409 |
| }, |
| { |
| "epoch": 44.06, |
| "learning_rate": 7.125e-05, |
| "loss": 0.2505, |
| "step": 1410 |
| }, |
| { |
| "epoch": 44.09, |
| "learning_rate": 7.087499999999999e-05, |
| "loss": 0.2029, |
| "step": 1411 |
| }, |
| { |
| "epoch": 44.12, |
| "learning_rate": 7.049999999999999e-05, |
| "loss": 0.2236, |
| "step": 1412 |
| }, |
| { |
| "epoch": 44.16, |
| "learning_rate": 7.0125e-05, |
| "loss": 0.2773, |
| "step": 1413 |
| }, |
| { |
| "epoch": 44.19, |
| "learning_rate": 6.975e-05, |
| "loss": 0.2362, |
| "step": 1414 |
| }, |
| { |
| "epoch": 44.22, |
| "learning_rate": 6.937499999999999e-05, |
| "loss": 0.2591, |
| "step": 1415 |
| }, |
| { |
| "epoch": 44.25, |
| "learning_rate": 6.9e-05, |
| "loss": 0.3038, |
| "step": 1416 |
| }, |
| { |
| "epoch": 44.28, |
| "learning_rate": 6.8625e-05, |
| "loss": 0.2447, |
| "step": 1417 |
| }, |
| { |
| "epoch": 44.31, |
| "learning_rate": 6.824999999999999e-05, |
| "loss": 0.2637, |
| "step": 1418 |
| }, |
| { |
| "epoch": 44.34, |
| "learning_rate": 6.7875e-05, |
| "loss": 0.2357, |
| "step": 1419 |
| }, |
| { |
| "epoch": 44.38, |
| "learning_rate": 6.75e-05, |
| "loss": 0.228, |
| "step": 1420 |
| }, |
| { |
| "epoch": 44.38, |
| "eval_loss": 1.1583712100982666, |
| "eval_mean_accuracy": 0.5388642753404104, |
| "eval_mean_iou": 0.4156287128541895, |
| "eval_overall_accuracy": 0.7080377946820176, |
| "eval_per_category_accuracy": [ |
| 0.7075756167421354, |
| 0.802308406362694, |
| 0.7593988652677361, |
| 0.3550981603181102, |
| 0.5587542316053055, |
| 0.5830463201216525, |
| 0.6073866132198902, |
| 0.5147598174670608, |
| 0.3446935751488425, |
| 0.607196558185965, |
| 0.19068417846856414, |
| 0.43546896117696826 |
| ], |
| "eval_per_category_iou": [ |
| 0.6065913780653972, |
| 0.5924635201772961, |
| 0.6040944916922074, |
| 0.28359321526648174, |
| 0.40200884170204954, |
| 0.41587367331089825, |
| 0.4682726044000129, |
| 0.4435354994760821, |
| 0.23528841727838748, |
| 0.43694121073415837, |
| 0.15987179256879344, |
| 0.33900990957850957 |
| ], |
| "eval_runtime": 123.731, |
| "eval_samples_per_second": 0.921, |
| "eval_steps_per_second": 0.081, |
| "step": 1420 |
| }, |
| { |
| "epoch": 44.41, |
| "learning_rate": 6.712499999999999e-05, |
| "loss": 0.2659, |
| "step": 1421 |
| }, |
| { |
| "epoch": 44.44, |
| "learning_rate": 6.675e-05, |
| "loss": 0.2424, |
| "step": 1422 |
| }, |
| { |
| "epoch": 44.47, |
| "learning_rate": 6.6375e-05, |
| "loss": 0.2496, |
| "step": 1423 |
| }, |
| { |
| "epoch": 44.5, |
| "learning_rate": 6.599999999999999e-05, |
| "loss": 0.2268, |
| "step": 1424 |
| }, |
| { |
| "epoch": 44.53, |
| "learning_rate": 6.5625e-05, |
| "loss": 0.2359, |
| "step": 1425 |
| }, |
| { |
| "epoch": 44.56, |
| "learning_rate": 6.525e-05, |
| "loss": 0.2591, |
| "step": 1426 |
| }, |
| { |
| "epoch": 44.59, |
| "learning_rate": 6.487499999999999e-05, |
| "loss": 0.2138, |
| "step": 1427 |
| }, |
| { |
| "epoch": 44.62, |
| "learning_rate": 6.45e-05, |
| "loss": 0.1963, |
| "step": 1428 |
| }, |
| { |
| "epoch": 44.66, |
| "learning_rate": 6.412499999999999e-05, |
| "loss": 0.2455, |
| "step": 1429 |
| }, |
| { |
| "epoch": 44.69, |
| "learning_rate": 6.374999999999999e-05, |
| "loss": 0.215, |
| "step": 1430 |
| }, |
| { |
| "epoch": 44.72, |
| "learning_rate": 6.3375e-05, |
| "loss": 0.2598, |
| "step": 1431 |
| }, |
| { |
| "epoch": 44.75, |
| "learning_rate": 6.299999999999999e-05, |
| "loss": 0.245, |
| "step": 1432 |
| }, |
| { |
| "epoch": 44.78, |
| "learning_rate": 6.262499999999999e-05, |
| "loss": 0.239, |
| "step": 1433 |
| }, |
| { |
| "epoch": 44.81, |
| "learning_rate": 6.225e-05, |
| "loss": 0.2267, |
| "step": 1434 |
| }, |
| { |
| "epoch": 44.84, |
| "learning_rate": 6.187499999999999e-05, |
| "loss": 0.2866, |
| "step": 1435 |
| }, |
| { |
| "epoch": 44.88, |
| "learning_rate": 6.149999999999999e-05, |
| "loss": 0.2084, |
| "step": 1436 |
| }, |
| { |
| "epoch": 44.91, |
| "learning_rate": 6.1125e-05, |
| "loss": 0.2597, |
| "step": 1437 |
| }, |
| { |
| "epoch": 44.94, |
| "learning_rate": 6.075e-05, |
| "loss": 0.2151, |
| "step": 1438 |
| }, |
| { |
| "epoch": 44.97, |
| "learning_rate": 6.0375e-05, |
| "loss": 0.287, |
| "step": 1439 |
| }, |
| { |
| "epoch": 45.0, |
| "learning_rate": 5.9999999999999995e-05, |
| "loss": 0.2298, |
| "step": 1440 |
| }, |
| { |
| "epoch": 45.0, |
| "eval_loss": 1.1716525554656982, |
| "eval_mean_accuracy": 0.5379453595950269, |
| "eval_mean_iou": 0.41453145156483057, |
| "eval_overall_accuracy": 0.7053226002475672, |
| "eval_per_category_accuracy": [ |
| 0.6825056434568719, |
| 0.8150425095313159, |
| 0.7440366346796163, |
| 0.36192670418664685, |
| 0.5600933536193774, |
| 0.542135596049924, |
| 0.6248675507636791, |
| 0.5191474224561802, |
| 0.3747770689333913, |
| 0.591828038642697, |
| 0.2171477718976594, |
| 0.42183602092296263 |
| ], |
| "eval_per_category_iou": [ |
| 0.5948243117826617, |
| 0.5926740583535665, |
| 0.6009514529005231, |
| 0.2857635987562155, |
| 0.3994965579825327, |
| 0.40232838433828233, |
| 0.47200192661614254, |
| 0.4451530987240548, |
| 0.2417762946066345, |
| 0.4330689308798251, |
| 0.17467942404664355, |
| 0.3316593797908842 |
| ], |
| "eval_runtime": 124.5862, |
| "eval_samples_per_second": 0.915, |
| "eval_steps_per_second": 0.08, |
| "step": 1440 |
| }, |
| { |
| "epoch": 45.03, |
| "learning_rate": 5.9625e-05, |
| "loss": 0.2074, |
| "step": 1441 |
| }, |
| { |
| "epoch": 45.06, |
| "learning_rate": 5.925e-05, |
| "loss": 0.2415, |
| "step": 1442 |
| }, |
| { |
| "epoch": 45.09, |
| "learning_rate": 5.8874999999999995e-05, |
| "loss": 0.2662, |
| "step": 1443 |
| }, |
| { |
| "epoch": 45.12, |
| "learning_rate": 5.85e-05, |
| "loss": 0.216, |
| "step": 1444 |
| }, |
| { |
| "epoch": 45.16, |
| "learning_rate": 5.8124999999999997e-05, |
| "loss": 0.207, |
| "step": 1445 |
| }, |
| { |
| "epoch": 45.19, |
| "learning_rate": 5.7749999999999994e-05, |
| "loss": 0.2531, |
| "step": 1446 |
| }, |
| { |
| "epoch": 45.22, |
| "learning_rate": 5.7375e-05, |
| "loss": 0.2615, |
| "step": 1447 |
| }, |
| { |
| "epoch": 45.25, |
| "learning_rate": 5.6999999999999996e-05, |
| "loss": 0.2134, |
| "step": 1448 |
| }, |
| { |
| "epoch": 45.28, |
| "learning_rate": 5.6624999999999994e-05, |
| "loss": 0.2598, |
| "step": 1449 |
| }, |
| { |
| "epoch": 45.31, |
| "learning_rate": 5.625e-05, |
| "loss": 0.2568, |
| "step": 1450 |
| }, |
| { |
| "epoch": 45.34, |
| "learning_rate": 5.5874999999999996e-05, |
| "loss": 0.2287, |
| "step": 1451 |
| }, |
| { |
| "epoch": 45.38, |
| "learning_rate": 5.5499999999999994e-05, |
| "loss": 0.2228, |
| "step": 1452 |
| }, |
| { |
| "epoch": 45.41, |
| "learning_rate": 5.512499999999999e-05, |
| "loss": 0.2221, |
| "step": 1453 |
| }, |
| { |
| "epoch": 45.44, |
| "learning_rate": 5.4749999999999996e-05, |
| "loss": 0.2452, |
| "step": 1454 |
| }, |
| { |
| "epoch": 45.47, |
| "learning_rate": 5.4374999999999994e-05, |
| "loss": 0.2364, |
| "step": 1455 |
| }, |
| { |
| "epoch": 45.5, |
| "learning_rate": 5.399999999999999e-05, |
| "loss": 0.2342, |
| "step": 1456 |
| }, |
| { |
| "epoch": 45.53, |
| "learning_rate": 5.3624999999999996e-05, |
| "loss": 0.2048, |
| "step": 1457 |
| }, |
| { |
| "epoch": 45.56, |
| "learning_rate": 5.324999999999999e-05, |
| "loss": 0.2523, |
| "step": 1458 |
| }, |
| { |
| "epoch": 45.59, |
| "learning_rate": 5.287499999999999e-05, |
| "loss": 0.2335, |
| "step": 1459 |
| }, |
| { |
| "epoch": 45.62, |
| "learning_rate": 5.2499999999999995e-05, |
| "loss": 0.2131, |
| "step": 1460 |
| }, |
| { |
| "epoch": 45.62, |
| "eval_loss": 1.1651273965835571, |
| "eval_mean_accuracy": 0.5385710717175287, |
| "eval_mean_iou": 0.41477686211852943, |
| "eval_overall_accuracy": 0.7067017471581175, |
| "eval_per_category_accuracy": [ |
| 0.7028630567631983, |
| 0.8068977800227759, |
| 0.7513259611595465, |
| 0.36922713929813117, |
| 0.567376475617242, |
| 0.5730931011580938, |
| 0.6197041953935218, |
| 0.5095600501869152, |
| 0.3600214723654206, |
| 0.581253086925206, |
| 0.1947825767569702, |
| 0.4267479649633229 |
| ], |
| "eval_per_category_iou": [ |
| 0.6027280222990506, |
| 0.5919536961027094, |
| 0.6032660982212309, |
| 0.28713515147099694, |
| 0.4005300146161613, |
| 0.41143265833602566, |
| 0.4703060000967844, |
| 0.44494737758001424, |
| 0.23920178625380425, |
| 0.4297041155468503, |
| 0.16184638860630723, |
| 0.33427103629241783 |
| ], |
| "eval_runtime": 124.9362, |
| "eval_samples_per_second": 0.912, |
| "eval_steps_per_second": 0.08, |
| "step": 1460 |
| }, |
| { |
| "epoch": 45.66, |
| "learning_rate": 5.212499999999999e-05, |
| "loss": 0.2332, |
| "step": 1461 |
| }, |
| { |
| "epoch": 45.69, |
| "learning_rate": 5.174999999999999e-05, |
| "loss": 0.2245, |
| "step": 1462 |
| }, |
| { |
| "epoch": 45.72, |
| "learning_rate": 5.1375e-05, |
| "loss": 0.2247, |
| "step": 1463 |
| }, |
| { |
| "epoch": 45.75, |
| "learning_rate": 5.1e-05, |
| "loss": 0.2351, |
| "step": 1464 |
| }, |
| { |
| "epoch": 45.78, |
| "learning_rate": 5.0625e-05, |
| "loss": 0.2105, |
| "step": 1465 |
| }, |
| { |
| "epoch": 45.81, |
| "learning_rate": 5.025e-05, |
| "loss": 0.2405, |
| "step": 1466 |
| }, |
| { |
| "epoch": 45.84, |
| "learning_rate": 4.9875e-05, |
| "loss": 0.3312, |
| "step": 1467 |
| }, |
| { |
| "epoch": 45.88, |
| "learning_rate": 4.95e-05, |
| "loss": 0.2472, |
| "step": 1468 |
| }, |
| { |
| "epoch": 45.91, |
| "learning_rate": 4.9124999999999995e-05, |
| "loss": 0.3075, |
| "step": 1469 |
| }, |
| { |
| "epoch": 45.94, |
| "learning_rate": 4.875e-05, |
| "loss": 0.2313, |
| "step": 1470 |
| }, |
| { |
| "epoch": 45.97, |
| "learning_rate": 4.8375e-05, |
| "loss": 0.232, |
| "step": 1471 |
| }, |
| { |
| "epoch": 46.0, |
| "learning_rate": 4.7999999999999994e-05, |
| "loss": 0.2113, |
| "step": 1472 |
| }, |
| { |
| "epoch": 46.03, |
| "learning_rate": 4.7625e-05, |
| "loss": 0.2275, |
| "step": 1473 |
| }, |
| { |
| "epoch": 46.06, |
| "learning_rate": 4.7249999999999997e-05, |
| "loss": 0.1768, |
| "step": 1474 |
| }, |
| { |
| "epoch": 46.09, |
| "learning_rate": 4.6874999999999994e-05, |
| "loss": 0.2636, |
| "step": 1475 |
| }, |
| { |
| "epoch": 46.12, |
| "learning_rate": 4.65e-05, |
| "loss": 0.2145, |
| "step": 1476 |
| }, |
| { |
| "epoch": 46.16, |
| "learning_rate": 4.6124999999999996e-05, |
| "loss": 0.2373, |
| "step": 1477 |
| }, |
| { |
| "epoch": 46.19, |
| "learning_rate": 4.5749999999999994e-05, |
| "loss": 0.2245, |
| "step": 1478 |
| }, |
| { |
| "epoch": 46.22, |
| "learning_rate": 4.537499999999999e-05, |
| "loss": 0.2209, |
| "step": 1479 |
| }, |
| { |
| "epoch": 46.25, |
| "learning_rate": 4.4999999999999996e-05, |
| "loss": 0.2236, |
| "step": 1480 |
| }, |
| { |
| "epoch": 46.25, |
| "eval_loss": 1.1631407737731934, |
| "eval_mean_accuracy": 0.5388222818505555, |
| "eval_mean_iou": 0.41579736978967735, |
| "eval_overall_accuracy": 0.7066818705776281, |
| "eval_per_category_accuracy": [ |
| 0.6890048147611364, |
| 0.8134493621508735, |
| 0.7522002808438099, |
| 0.3676761438157959, |
| 0.5591776453737487, |
| 0.5621853893641191, |
| 0.6024468274373415, |
| 0.5226682209895737, |
| 0.3634264394026778, |
| 0.5972996217196449, |
| 0.22089111850462026, |
| 0.4154415178433254 |
| ], |
| "eval_per_category_iou": [ |
| 0.5989571778368314, |
| 0.5934079057996247, |
| 0.6027309057700777, |
| 0.28637950791317135, |
| 0.40160634389436545, |
| 0.40849887790972117, |
| 0.46730974497509836, |
| 0.44940164168516866, |
| 0.2401122615723948, |
| 0.434476583943655, |
| 0.17743160813990366, |
| 0.32925587803611656 |
| ], |
| "eval_runtime": 124.7245, |
| "eval_samples_per_second": 0.914, |
| "eval_steps_per_second": 0.08, |
| "step": 1480 |
| }, |
| { |
| "epoch": 46.28, |
| "learning_rate": 4.4624999999999994e-05, |
| "loss": 0.2446, |
| "step": 1481 |
| }, |
| { |
| "epoch": 46.31, |
| "learning_rate": 4.424999999999999e-05, |
| "loss": 0.2809, |
| "step": 1482 |
| }, |
| { |
| "epoch": 46.34, |
| "learning_rate": 4.3874999999999996e-05, |
| "loss": 0.2417, |
| "step": 1483 |
| }, |
| { |
| "epoch": 46.38, |
| "learning_rate": 4.3499999999999993e-05, |
| "loss": 0.2334, |
| "step": 1484 |
| }, |
| { |
| "epoch": 46.41, |
| "learning_rate": 4.312499999999999e-05, |
| "loss": 0.2126, |
| "step": 1485 |
| }, |
| { |
| "epoch": 46.44, |
| "learning_rate": 4.2749999999999996e-05, |
| "loss": 0.2799, |
| "step": 1486 |
| }, |
| { |
| "epoch": 46.47, |
| "learning_rate": 4.237499999999999e-05, |
| "loss": 0.1999, |
| "step": 1487 |
| }, |
| { |
| "epoch": 46.5, |
| "learning_rate": 4.2e-05, |
| "loss": 0.2346, |
| "step": 1488 |
| }, |
| { |
| "epoch": 46.53, |
| "learning_rate": 4.1625e-05, |
| "loss": 0.2567, |
| "step": 1489 |
| }, |
| { |
| "epoch": 46.56, |
| "learning_rate": 4.125e-05, |
| "loss": 0.2486, |
| "step": 1490 |
| }, |
| { |
| "epoch": 46.59, |
| "learning_rate": 4.0875e-05, |
| "loss": 0.2177, |
| "step": 1491 |
| }, |
| { |
| "epoch": 46.62, |
| "learning_rate": 4.05e-05, |
| "loss": 0.2345, |
| "step": 1492 |
| }, |
| { |
| "epoch": 46.66, |
| "learning_rate": 4.0125e-05, |
| "loss": 0.2066, |
| "step": 1493 |
| }, |
| { |
| "epoch": 46.69, |
| "learning_rate": 3.975e-05, |
| "loss": 0.201, |
| "step": 1494 |
| }, |
| { |
| "epoch": 46.72, |
| "learning_rate": 3.9374999999999995e-05, |
| "loss": 0.2173, |
| "step": 1495 |
| }, |
| { |
| "epoch": 46.75, |
| "learning_rate": 3.9e-05, |
| "loss": 0.2462, |
| "step": 1496 |
| }, |
| { |
| "epoch": 46.78, |
| "learning_rate": 3.8625e-05, |
| "loss": 0.2917, |
| "step": 1497 |
| }, |
| { |
| "epoch": 46.81, |
| "learning_rate": 3.8249999999999995e-05, |
| "loss": 0.2307, |
| "step": 1498 |
| }, |
| { |
| "epoch": 46.84, |
| "learning_rate": 3.7875e-05, |
| "loss": 0.311, |
| "step": 1499 |
| }, |
| { |
| "epoch": 46.88, |
| "learning_rate": 3.75e-05, |
| "loss": 0.2541, |
| "step": 1500 |
| }, |
| { |
| "epoch": 46.88, |
| "eval_loss": 1.1742721796035767, |
| "eval_mean_accuracy": 0.5370777663952354, |
| "eval_mean_iou": 0.4143455278141383, |
| "eval_overall_accuracy": 0.7066309075606497, |
| "eval_per_category_accuracy": [ |
| 0.6920560185688931, |
| 0.8121339201013871, |
| 0.7566460708832426, |
| 0.37307166501273414, |
| 0.5658712295409101, |
| 0.5486101785954075, |
| 0.6077739239139163, |
| 0.5225685757480626, |
| 0.360737957285964, |
| 0.5795345790385849, |
| 0.19024036386675766, |
| 0.4356887141869655 |
| ], |
| "eval_per_category_iou": [ |
| 0.6000424730412047, |
| 0.59315042083966, |
| 0.6026233782289969, |
| 0.28847856533427774, |
| 0.40007276185674506, |
| 0.4059388259305011, |
| 0.46811918984664713, |
| 0.44951722790382165, |
| 0.2374007880027386, |
| 0.4298806507147417, |
| 0.15971815350457924, |
| 0.33720389856574556 |
| ], |
| "eval_runtime": 125.1267, |
| "eval_samples_per_second": 0.911, |
| "eval_steps_per_second": 0.08, |
| "step": 1500 |
| }, |
| { |
| "epoch": 46.91, |
| "learning_rate": 3.7124999999999994e-05, |
| "loss": 0.2471, |
| "step": 1501 |
| }, |
| { |
| "epoch": 46.94, |
| "learning_rate": 3.675e-05, |
| "loss": 0.2114, |
| "step": 1502 |
| }, |
| { |
| "epoch": 46.97, |
| "learning_rate": 3.6374999999999996e-05, |
| "loss": 0.2604, |
| "step": 1503 |
| }, |
| { |
| "epoch": 47.0, |
| "learning_rate": 3.5999999999999994e-05, |
| "loss": 0.2488, |
| "step": 1504 |
| }, |
| { |
| "epoch": 47.03, |
| "learning_rate": 3.5625e-05, |
| "loss": 0.2071, |
| "step": 1505 |
| }, |
| { |
| "epoch": 47.06, |
| "learning_rate": 3.5249999999999996e-05, |
| "loss": 0.2285, |
| "step": 1506 |
| }, |
| { |
| "epoch": 47.09, |
| "learning_rate": 3.4875e-05, |
| "loss": 0.2856, |
| "step": 1507 |
| }, |
| { |
| "epoch": 47.12, |
| "learning_rate": 3.45e-05, |
| "loss": 0.2539, |
| "step": 1508 |
| }, |
| { |
| "epoch": 47.16, |
| "learning_rate": 3.4124999999999996e-05, |
| "loss": 0.2393, |
| "step": 1509 |
| }, |
| { |
| "epoch": 47.19, |
| "learning_rate": 3.375e-05, |
| "loss": 0.1994, |
| "step": 1510 |
| }, |
| { |
| "epoch": 47.22, |
| "learning_rate": 3.3375e-05, |
| "loss": 0.2193, |
| "step": 1511 |
| }, |
| { |
| "epoch": 47.25, |
| "learning_rate": 3.2999999999999996e-05, |
| "loss": 0.2737, |
| "step": 1512 |
| }, |
| { |
| "epoch": 47.28, |
| "learning_rate": 3.2625e-05, |
| "loss": 0.231, |
| "step": 1513 |
| }, |
| { |
| "epoch": 47.31, |
| "learning_rate": 3.225e-05, |
| "loss": 0.2409, |
| "step": 1514 |
| }, |
| { |
| "epoch": 47.34, |
| "learning_rate": 3.1874999999999996e-05, |
| "loss": 0.2966, |
| "step": 1515 |
| }, |
| { |
| "epoch": 47.38, |
| "learning_rate": 3.149999999999999e-05, |
| "loss": 0.2489, |
| "step": 1516 |
| }, |
| { |
| "epoch": 47.41, |
| "learning_rate": 3.1125e-05, |
| "loss": 0.2269, |
| "step": 1517 |
| }, |
| { |
| "epoch": 47.44, |
| "learning_rate": 3.0749999999999995e-05, |
| "loss": 0.2191, |
| "step": 1518 |
| }, |
| { |
| "epoch": 47.47, |
| "learning_rate": 3.0375e-05, |
| "loss": 0.2428, |
| "step": 1519 |
| }, |
| { |
| "epoch": 47.5, |
| "learning_rate": 2.9999999999999997e-05, |
| "loss": 0.2565, |
| "step": 1520 |
| }, |
| { |
| "epoch": 47.5, |
| "eval_loss": 1.165220856666565, |
| "eval_mean_accuracy": 0.5421998727049119, |
| "eval_mean_iou": 0.4163862285048401, |
| "eval_overall_accuracy": 0.7065070972108004, |
| "eval_per_category_accuracy": [ |
| 0.6998926153643288, |
| 0.8057150867806927, |
| 0.7504088386186776, |
| 0.3730022692417124, |
| 0.5614365274398458, |
| 0.5652248864951966, |
| 0.6063955071959493, |
| 0.523395738398026, |
| 0.3746351028810236, |
| 0.5963316245610503, |
| 0.22688414602418636, |
| 0.42307612945825207 |
| ], |
| "eval_per_category_iou": [ |
| 0.6028865669550012, |
| 0.5920776237505179, |
| 0.6023657478955948, |
| 0.2876354227705579, |
| 0.3989585837660571, |
| 0.40995231658732595, |
| 0.4668406832828899, |
| 0.4494173107513457, |
| 0.24100804121210392, |
| 0.43316164208360136, |
| 0.180027152464244, |
| 0.3323036505388412 |
| ], |
| "eval_runtime": 125.0757, |
| "eval_samples_per_second": 0.911, |
| "eval_steps_per_second": 0.08, |
| "step": 1520 |
| }, |
| { |
| "epoch": 47.53, |
| "learning_rate": 2.9625e-05, |
| "loss": 0.2388, |
| "step": 1521 |
| }, |
| { |
| "epoch": 47.56, |
| "learning_rate": 2.925e-05, |
| "loss": 0.2216, |
| "step": 1522 |
| }, |
| { |
| "epoch": 47.59, |
| "learning_rate": 2.8874999999999997e-05, |
| "loss": 0.2088, |
| "step": 1523 |
| }, |
| { |
| "epoch": 47.62, |
| "learning_rate": 2.8499999999999998e-05, |
| "loss": 0.2318, |
| "step": 1524 |
| }, |
| { |
| "epoch": 47.66, |
| "learning_rate": 2.8125e-05, |
| "loss": 0.2048, |
| "step": 1525 |
| }, |
| { |
| "epoch": 47.69, |
| "learning_rate": 2.7749999999999997e-05, |
| "loss": 0.1987, |
| "step": 1526 |
| }, |
| { |
| "epoch": 47.72, |
| "learning_rate": 2.7374999999999998e-05, |
| "loss": 0.2493, |
| "step": 1527 |
| }, |
| { |
| "epoch": 47.75, |
| "learning_rate": 2.6999999999999996e-05, |
| "loss": 0.2576, |
| "step": 1528 |
| }, |
| { |
| "epoch": 47.78, |
| "learning_rate": 2.6624999999999997e-05, |
| "loss": 0.27, |
| "step": 1529 |
| }, |
| { |
| "epoch": 47.81, |
| "learning_rate": 2.6249999999999998e-05, |
| "loss": 0.2654, |
| "step": 1530 |
| }, |
| { |
| "epoch": 47.84, |
| "learning_rate": 2.5874999999999995e-05, |
| "loss": 0.235, |
| "step": 1531 |
| }, |
| { |
| "epoch": 47.88, |
| "learning_rate": 2.55e-05, |
| "loss": 0.2739, |
| "step": 1532 |
| }, |
| { |
| "epoch": 47.91, |
| "learning_rate": 2.5125e-05, |
| "loss": 0.2274, |
| "step": 1533 |
| }, |
| { |
| "epoch": 47.94, |
| "learning_rate": 2.475e-05, |
| "loss": 0.2517, |
| "step": 1534 |
| }, |
| { |
| "epoch": 47.97, |
| "learning_rate": 2.4375e-05, |
| "loss": 0.2861, |
| "step": 1535 |
| }, |
| { |
| "epoch": 48.0, |
| "learning_rate": 2.3999999999999997e-05, |
| "loss": 0.2618, |
| "step": 1536 |
| }, |
| { |
| "epoch": 48.03, |
| "learning_rate": 2.3624999999999998e-05, |
| "loss": 0.2497, |
| "step": 1537 |
| }, |
| { |
| "epoch": 48.06, |
| "learning_rate": 2.325e-05, |
| "loss": 0.2813, |
| "step": 1538 |
| }, |
| { |
| "epoch": 48.09, |
| "learning_rate": 2.2874999999999997e-05, |
| "loss": 0.2237, |
| "step": 1539 |
| }, |
| { |
| "epoch": 48.12, |
| "learning_rate": 2.2499999999999998e-05, |
| "loss": 0.2488, |
| "step": 1540 |
| }, |
| { |
| "epoch": 48.12, |
| "eval_loss": 1.1770268678665161, |
| "eval_mean_accuracy": 0.5357019382168873, |
| "eval_mean_iou": 0.4138841161101329, |
| "eval_overall_accuracy": 0.7071278555351391, |
| "eval_per_category_accuracy": [ |
| 0.7004789504528495, |
| 0.8122536527346174, |
| 0.7540579248204919, |
| 0.37313759099520477, |
| 0.5595787742070109, |
| 0.5727065860005298, |
| 0.595211517297516, |
| 0.5122515431619257, |
| 0.3549262220171601, |
| 0.5823415621784317, |
| 0.20312935206128926, |
| 0.4083495826756219 |
| ], |
| "eval_per_category_iou": [ |
| 0.6035941856023294, |
| 0.5934513401985494, |
| 0.6022197731275365, |
| 0.2855401788559169, |
| 0.3986554262844244, |
| 0.40997269498663325, |
| 0.46628468245789634, |
| 0.4455256399264928, |
| 0.23833356421175872, |
| 0.42950408266968554, |
| 0.16799181873640992, |
| 0.3255360062639615 |
| ], |
| "eval_runtime": 125.8575, |
| "eval_samples_per_second": 0.906, |
| "eval_steps_per_second": 0.079, |
| "step": 1540 |
| }, |
| { |
| "epoch": 48.16, |
| "learning_rate": 2.2124999999999996e-05, |
| "loss": 0.2103, |
| "step": 1541 |
| }, |
| { |
| "epoch": 48.19, |
| "learning_rate": 2.1749999999999997e-05, |
| "loss": 0.2193, |
| "step": 1542 |
| }, |
| { |
| "epoch": 48.22, |
| "learning_rate": 2.1374999999999998e-05, |
| "loss": 0.2861, |
| "step": 1543 |
| }, |
| { |
| "epoch": 48.25, |
| "learning_rate": 2.1e-05, |
| "loss": 0.2207, |
| "step": 1544 |
| }, |
| { |
| "epoch": 48.28, |
| "learning_rate": 2.0625e-05, |
| "loss": 0.2066, |
| "step": 1545 |
| }, |
| { |
| "epoch": 48.31, |
| "learning_rate": 2.025e-05, |
| "loss": 0.2537, |
| "step": 1546 |
| }, |
| { |
| "epoch": 48.34, |
| "learning_rate": 1.9875e-05, |
| "loss": 0.2323, |
| "step": 1547 |
| }, |
| { |
| "epoch": 48.38, |
| "learning_rate": 1.95e-05, |
| "loss": 0.2573, |
| "step": 1548 |
| }, |
| { |
| "epoch": 48.41, |
| "learning_rate": 1.9124999999999997e-05, |
| "loss": 0.242, |
| "step": 1549 |
| }, |
| { |
| "epoch": 48.44, |
| "learning_rate": 1.875e-05, |
| "loss": 0.219, |
| "step": 1550 |
| }, |
| { |
| "epoch": 48.47, |
| "learning_rate": 1.8375e-05, |
| "loss": 0.2565, |
| "step": 1551 |
| }, |
| { |
| "epoch": 48.5, |
| "learning_rate": 1.7999999999999997e-05, |
| "loss": 0.2127, |
| "step": 1552 |
| }, |
| { |
| "epoch": 48.53, |
| "learning_rate": 1.7624999999999998e-05, |
| "loss": 0.2251, |
| "step": 1553 |
| }, |
| { |
| "epoch": 48.56, |
| "learning_rate": 1.725e-05, |
| "loss": 0.238, |
| "step": 1554 |
| }, |
| { |
| "epoch": 48.59, |
| "learning_rate": 1.6875e-05, |
| "loss": 0.1888, |
| "step": 1555 |
| }, |
| { |
| "epoch": 48.62, |
| "learning_rate": 1.6499999999999998e-05, |
| "loss": 0.2394, |
| "step": 1556 |
| }, |
| { |
| "epoch": 48.66, |
| "learning_rate": 1.6125e-05, |
| "loss": 0.2425, |
| "step": 1557 |
| }, |
| { |
| "epoch": 48.69, |
| "learning_rate": 1.5749999999999997e-05, |
| "loss": 0.2046, |
| "step": 1558 |
| }, |
| { |
| "epoch": 48.72, |
| "learning_rate": 1.5374999999999998e-05, |
| "loss": 0.2635, |
| "step": 1559 |
| }, |
| { |
| "epoch": 48.75, |
| "learning_rate": 1.4999999999999999e-05, |
| "loss": 0.2271, |
| "step": 1560 |
| }, |
| { |
| "epoch": 48.75, |
| "eval_loss": 1.1763782501220703, |
| "eval_mean_accuracy": 0.5379843252388352, |
| "eval_mean_iou": 0.41417587578295884, |
| "eval_overall_accuracy": 0.7066306398625959, |
| "eval_per_category_accuracy": [ |
| 0.6904259394760635, |
| 0.8116648527563487, |
| 0.7527524376940218, |
| 0.3743867148735956, |
| 0.5684562820219324, |
| 0.5636425333618834, |
| 0.6004071485588421, |
| 0.5293262374492532, |
| 0.3514413990754461, |
| 0.6006850040237598, |
| 0.1899495887828155, |
| 0.42267376479206 |
| ], |
| "eval_per_category_iou": [ |
| 0.5996213429826344, |
| 0.5934785532237628, |
| 0.6027337134689185, |
| 0.28731463507507793, |
| 0.3995881562158221, |
| 0.4078282927751562, |
| 0.46648158987447685, |
| 0.4517333669829548, |
| 0.2368072557477707, |
| 0.4340776570593164, |
| 0.1591440037953097, |
| 0.33130194219430675 |
| ], |
| "eval_runtime": 124.036, |
| "eval_samples_per_second": 0.919, |
| "eval_steps_per_second": 0.081, |
| "step": 1560 |
| }, |
| { |
| "epoch": 48.78, |
| "learning_rate": 1.4625e-05, |
| "loss": 0.2306, |
| "step": 1561 |
| }, |
| { |
| "epoch": 48.81, |
| "learning_rate": 1.4249999999999999e-05, |
| "loss": 0.2288, |
| "step": 1562 |
| }, |
| { |
| "epoch": 48.84, |
| "learning_rate": 1.3874999999999998e-05, |
| "loss": 0.2462, |
| "step": 1563 |
| }, |
| { |
| "epoch": 48.88, |
| "learning_rate": 1.3499999999999998e-05, |
| "loss": 0.2575, |
| "step": 1564 |
| }, |
| { |
| "epoch": 48.91, |
| "learning_rate": 1.3124999999999999e-05, |
| "loss": 0.2325, |
| "step": 1565 |
| }, |
| { |
| "epoch": 48.94, |
| "learning_rate": 1.275e-05, |
| "loss": 0.2485, |
| "step": 1566 |
| }, |
| { |
| "epoch": 48.97, |
| "learning_rate": 1.2375e-05, |
| "loss": 0.2532, |
| "step": 1567 |
| }, |
| { |
| "epoch": 49.0, |
| "learning_rate": 1.1999999999999999e-05, |
| "loss": 0.281, |
| "step": 1568 |
| }, |
| { |
| "epoch": 49.03, |
| "learning_rate": 1.1625e-05, |
| "loss": 0.2455, |
| "step": 1569 |
| }, |
| { |
| "epoch": 49.06, |
| "learning_rate": 1.1249999999999999e-05, |
| "loss": 0.2667, |
| "step": 1570 |
| }, |
| { |
| "epoch": 49.09, |
| "learning_rate": 1.0874999999999998e-05, |
| "loss": 0.2176, |
| "step": 1571 |
| }, |
| { |
| "epoch": 49.12, |
| "learning_rate": 1.05e-05, |
| "loss": 0.2248, |
| "step": 1572 |
| }, |
| { |
| "epoch": 49.16, |
| "learning_rate": 1.0125e-05, |
| "loss": 0.2083, |
| "step": 1573 |
| }, |
| { |
| "epoch": 49.19, |
| "learning_rate": 9.75e-06, |
| "loss": 0.2236, |
| "step": 1574 |
| }, |
| { |
| "epoch": 49.22, |
| "learning_rate": 9.375e-06, |
| "loss": 0.2599, |
| "step": 1575 |
| }, |
| { |
| "epoch": 49.25, |
| "learning_rate": 8.999999999999999e-06, |
| "loss": 0.234, |
| "step": 1576 |
| }, |
| { |
| "epoch": 49.28, |
| "learning_rate": 8.625e-06, |
| "loss": 0.2259, |
| "step": 1577 |
| }, |
| { |
| "epoch": 49.31, |
| "learning_rate": 8.249999999999999e-06, |
| "loss": 0.23, |
| "step": 1578 |
| }, |
| { |
| "epoch": 49.34, |
| "learning_rate": 7.874999999999998e-06, |
| "loss": 0.2236, |
| "step": 1579 |
| }, |
| { |
| "epoch": 49.38, |
| "learning_rate": 7.499999999999999e-06, |
| "loss": 0.2433, |
| "step": 1580 |
| }, |
| { |
| "epoch": 49.38, |
| "eval_loss": 1.1711814403533936, |
| "eval_mean_accuracy": 0.5400736480026452, |
| "eval_mean_iou": 0.41527787020587464, |
| "eval_overall_accuracy": 0.7070872323554859, |
| "eval_per_category_accuracy": [ |
| 0.698066357038275, |
| 0.8064249937708309, |
| 0.7526759632568865, |
| 0.3702125592466395, |
| 0.559372132080785, |
| 0.5698685122614222, |
| 0.6078101358893739, |
| 0.526311165302884, |
| 0.36216649068791423, |
| 0.6061143415350578, |
| 0.19281142776687796, |
| 0.4290496971947961 |
| ], |
| "eval_per_category_iou": [ |
| 0.6032635792650157, |
| 0.5929435291876964, |
| 0.6034415546088335, |
| 0.28745315577204406, |
| 0.40024352943734554, |
| 0.40993422349392167, |
| 0.46782541465655036, |
| 0.4509266628661968, |
| 0.23886184785137135, |
| 0.4348354802463671, |
| 0.16051348823935604, |
| 0.33309197684579644 |
| ], |
| "eval_runtime": 125.5903, |
| "eval_samples_per_second": 0.908, |
| "eval_steps_per_second": 0.08, |
| "step": 1580 |
| }, |
| { |
| "epoch": 49.41, |
| "learning_rate": 7.1249999999999995e-06, |
| "loss": 0.2603, |
| "step": 1581 |
| }, |
| { |
| "epoch": 49.44, |
| "learning_rate": 6.749999999999999e-06, |
| "loss": 0.2196, |
| "step": 1582 |
| }, |
| { |
| "epoch": 49.47, |
| "learning_rate": 6.375e-06, |
| "loss": 0.2607, |
| "step": 1583 |
| }, |
| { |
| "epoch": 49.5, |
| "learning_rate": 5.999999999999999e-06, |
| "loss": 0.2358, |
| "step": 1584 |
| }, |
| { |
| "epoch": 49.53, |
| "learning_rate": 5.6249999999999995e-06, |
| "loss": 0.2127, |
| "step": 1585 |
| }, |
| { |
| "epoch": 49.56, |
| "learning_rate": 5.25e-06, |
| "loss": 0.2152, |
| "step": 1586 |
| }, |
| { |
| "epoch": 49.59, |
| "learning_rate": 4.875e-06, |
| "loss": 0.2255, |
| "step": 1587 |
| }, |
| { |
| "epoch": 49.62, |
| "learning_rate": 4.499999999999999e-06, |
| "loss": 0.2264, |
| "step": 1588 |
| }, |
| { |
| "epoch": 49.66, |
| "learning_rate": 4.1249999999999995e-06, |
| "loss": 0.214, |
| "step": 1589 |
| }, |
| { |
| "epoch": 49.69, |
| "learning_rate": 3.7499999999999997e-06, |
| "loss": 0.2275, |
| "step": 1590 |
| }, |
| { |
| "epoch": 49.72, |
| "learning_rate": 3.3749999999999995e-06, |
| "loss": 0.2205, |
| "step": 1591 |
| }, |
| { |
| "epoch": 49.75, |
| "learning_rate": 2.9999999999999997e-06, |
| "loss": 0.2527, |
| "step": 1592 |
| }, |
| { |
| "epoch": 49.78, |
| "learning_rate": 2.625e-06, |
| "loss": 0.2436, |
| "step": 1593 |
| }, |
| { |
| "epoch": 49.81, |
| "learning_rate": 2.2499999999999996e-06, |
| "loss": 0.2245, |
| "step": 1594 |
| }, |
| { |
| "epoch": 49.84, |
| "learning_rate": 1.8749999999999998e-06, |
| "loss": 0.2298, |
| "step": 1595 |
| }, |
| { |
| "epoch": 49.88, |
| "learning_rate": 1.4999999999999998e-06, |
| "loss": 0.2435, |
| "step": 1596 |
| }, |
| { |
| "epoch": 49.91, |
| "learning_rate": 1.1249999999999998e-06, |
| "loss": 0.2179, |
| "step": 1597 |
| }, |
| { |
| "epoch": 49.94, |
| "learning_rate": 7.499999999999999e-07, |
| "loss": 0.2251, |
| "step": 1598 |
| }, |
| { |
| "epoch": 49.97, |
| "learning_rate": 3.7499999999999996e-07, |
| "loss": 0.2433, |
| "step": 1599 |
| }, |
| { |
| "epoch": 50.0, |
| "learning_rate": 0.0, |
| "loss": 0.2299, |
| "step": 1600 |
| }, |
| { |
| "epoch": 50.0, |
| "eval_loss": 1.1650828123092651, |
| "eval_mean_accuracy": 0.5417148369282497, |
| "eval_mean_iou": 0.4159080609633914, |
| "eval_overall_accuracy": 0.7073247474536561, |
| "eval_per_category_accuracy": [ |
| 0.7052524493403106, |
| 0.8016972929429861, |
| 0.7569536807460481, |
| 0.36770737191275565, |
| 0.5672630250381376, |
| 0.5649218440946654, |
| 0.6160593813420158, |
| 0.5216406964023782, |
| 0.363590587650728, |
| 0.5980214471870451, |
| 0.20086436719689757, |
| 0.43660589928502896 |
| ], |
| "eval_per_category_iou": [ |
| 0.6059737667522098, |
| 0.5922624046091037, |
| 0.6036323553739977, |
| 0.28742998955776455, |
| 0.3995807331319292, |
| 0.4092660539603954, |
| 0.4687620360847436, |
| 0.4498044099685319, |
| 0.23927425500048172, |
| 0.43375479764946384, |
| 0.16512549537648613, |
| 0.3360304340955885 |
| ], |
| "eval_runtime": 125.2298, |
| "eval_samples_per_second": 0.91, |
| "eval_steps_per_second": 0.08, |
| "step": 1600 |
| } |
| ], |
| "max_steps": 1600, |
| "num_train_epochs": 50, |
| "total_flos": 3.31530475143168e+17, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|