| { |
| "best_metric": 0.9254727474972191, |
| "best_model_checkpoint": "luna-train/412\\checkpoint-70000", |
| "epoch": 250.0, |
| "eval_steps": 2000, |
| "global_step": 100000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9996000000000003e-05, |
| "loss": 0.6265, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.9992e-05, |
| "loss": 0.5784, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.9988000000000002e-05, |
| "loss": 0.5799, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 1.9984e-05, |
| "loss": 0.5502, |
| "step": 800 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.9980020000000003e-05, |
| "loss": 0.5103, |
| "step": 1000 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 1.997602e-05, |
| "loss": 0.5694, |
| "step": 1200 |
| }, |
| { |
| "epoch": 3.5, |
| "learning_rate": 1.997202e-05, |
| "loss": 0.5522, |
| "step": 1400 |
| }, |
| { |
| "epoch": 4.0, |
| "learning_rate": 1.996802e-05, |
| "loss": 0.4973, |
| "step": 1600 |
| }, |
| { |
| "epoch": 4.5, |
| "learning_rate": 1.9964020000000002e-05, |
| "loss": 0.5362, |
| "step": 1800 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 1.9960020000000003e-05, |
| "loss": 0.5258, |
| "step": 2000 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_f1": 0.19816513761467888, |
| "eval_iou": 0.004486772697418928, |
| "eval_loss": 0.5519086718559265, |
| "eval_runtime": 18.9714, |
| "eval_samples_per_second": 4.639, |
| "eval_steps_per_second": 2.319, |
| "step": 2000 |
| }, |
| { |
| "epoch": 5.5, |
| "learning_rate": 1.995602e-05, |
| "loss": 0.5187, |
| "step": 2200 |
| }, |
| { |
| "epoch": 6.0, |
| "learning_rate": 1.995202e-05, |
| "loss": 0.5193, |
| "step": 2400 |
| }, |
| { |
| "epoch": 6.5, |
| "learning_rate": 1.994802e-05, |
| "loss": 0.5018, |
| "step": 2600 |
| }, |
| { |
| "epoch": 7.0, |
| "learning_rate": 1.9944020000000002e-05, |
| "loss": 0.5123, |
| "step": 2800 |
| }, |
| { |
| "epoch": 7.5, |
| "learning_rate": 1.9940020000000004e-05, |
| "loss": 0.5044, |
| "step": 3000 |
| }, |
| { |
| "epoch": 8.0, |
| "learning_rate": 1.993604e-05, |
| "loss": 0.4726, |
| "step": 3200 |
| }, |
| { |
| "epoch": 8.5, |
| "learning_rate": 1.993204e-05, |
| "loss": 0.4859, |
| "step": 3400 |
| }, |
| { |
| "epoch": 9.0, |
| "learning_rate": 1.9928040000000002e-05, |
| "loss": 0.4752, |
| "step": 3600 |
| }, |
| { |
| "epoch": 9.5, |
| "learning_rate": 1.9924040000000004e-05, |
| "loss": 0.5021, |
| "step": 3800 |
| }, |
| { |
| "epoch": 10.0, |
| "learning_rate": 1.9920040000000002e-05, |
| "loss": 0.4533, |
| "step": 4000 |
| }, |
| { |
| "epoch": 10.0, |
| "eval_f1": 0.4993757802746566, |
| "eval_iou": 0.006175489164888859, |
| "eval_loss": 0.5687549114227295, |
| "eval_runtime": 19.1043, |
| "eval_samples_per_second": 4.606, |
| "eval_steps_per_second": 2.303, |
| "step": 4000 |
| }, |
| { |
| "epoch": 10.5, |
| "learning_rate": 1.991604e-05, |
| "loss": 0.5005, |
| "step": 4200 |
| }, |
| { |
| "epoch": 11.0, |
| "learning_rate": 1.991204e-05, |
| "loss": 0.4481, |
| "step": 4400 |
| }, |
| { |
| "epoch": 11.5, |
| "learning_rate": 1.9908040000000003e-05, |
| "loss": 0.4525, |
| "step": 4600 |
| }, |
| { |
| "epoch": 12.0, |
| "learning_rate": 1.990404e-05, |
| "loss": 0.4571, |
| "step": 4800 |
| }, |
| { |
| "epoch": 12.5, |
| "learning_rate": 1.990006e-05, |
| "loss": 0.4579, |
| "step": 5000 |
| }, |
| { |
| "epoch": 13.0, |
| "learning_rate": 1.989606e-05, |
| "loss": 0.4757, |
| "step": 5200 |
| }, |
| { |
| "epoch": 13.5, |
| "learning_rate": 1.9892060000000003e-05, |
| "loss": 0.443, |
| "step": 5400 |
| }, |
| { |
| "epoch": 14.0, |
| "learning_rate": 1.988806e-05, |
| "loss": 0.4815, |
| "step": 5600 |
| }, |
| { |
| "epoch": 14.5, |
| "learning_rate": 1.9884060000000002e-05, |
| "loss": 0.4158, |
| "step": 5800 |
| }, |
| { |
| "epoch": 15.0, |
| "learning_rate": 1.988006e-05, |
| "loss": 0.4632, |
| "step": 6000 |
| }, |
| { |
| "epoch": 15.0, |
| "eval_f1": 0.5548549810844893, |
| "eval_iou": 0.006495922803878784, |
| "eval_loss": 0.5439079403877258, |
| "eval_runtime": 18.9599, |
| "eval_samples_per_second": 4.641, |
| "eval_steps_per_second": 2.321, |
| "step": 6000 |
| }, |
| { |
| "epoch": 15.5, |
| "learning_rate": 1.987606e-05, |
| "loss": 0.4446, |
| "step": 6200 |
| }, |
| { |
| "epoch": 16.0, |
| "learning_rate": 1.9872060000000003e-05, |
| "loss": 0.4299, |
| "step": 6400 |
| }, |
| { |
| "epoch": 16.5, |
| "learning_rate": 1.986806e-05, |
| "loss": 0.4457, |
| "step": 6600 |
| }, |
| { |
| "epoch": 17.0, |
| "learning_rate": 1.986408e-05, |
| "loss": 0.4282, |
| "step": 6800 |
| }, |
| { |
| "epoch": 17.5, |
| "learning_rate": 1.9860080000000002e-05, |
| "loss": 0.4379, |
| "step": 7000 |
| }, |
| { |
| "epoch": 18.0, |
| "learning_rate": 1.9856080000000003e-05, |
| "loss": 0.4092, |
| "step": 7200 |
| }, |
| { |
| "epoch": 18.5, |
| "learning_rate": 1.985208e-05, |
| "loss": 0.3935, |
| "step": 7400 |
| }, |
| { |
| "epoch": 19.0, |
| "learning_rate": 1.984808e-05, |
| "loss": 0.4032, |
| "step": 7600 |
| }, |
| { |
| "epoch": 19.5, |
| "learning_rate": 1.984408e-05, |
| "loss": 0.3619, |
| "step": 7800 |
| }, |
| { |
| "epoch": 20.0, |
| "learning_rate": 1.9840080000000002e-05, |
| "loss": 0.3754, |
| "step": 8000 |
| }, |
| { |
| "epoch": 20.0, |
| "eval_f1": 0.6920565832426551, |
| "eval_iou": 0.006737241987138987, |
| "eval_loss": 0.49375441670417786, |
| "eval_runtime": 19.0741, |
| "eval_samples_per_second": 4.614, |
| "eval_steps_per_second": 2.307, |
| "step": 8000 |
| }, |
| { |
| "epoch": 20.5, |
| "learning_rate": 1.9836080000000003e-05, |
| "loss": 0.3708, |
| "step": 8200 |
| }, |
| { |
| "epoch": 21.0, |
| "learning_rate": 1.983208e-05, |
| "loss": 0.3571, |
| "step": 8400 |
| }, |
| { |
| "epoch": 21.5, |
| "learning_rate": 1.982808e-05, |
| "loss": 0.3529, |
| "step": 8600 |
| }, |
| { |
| "epoch": 22.0, |
| "learning_rate": 1.982408e-05, |
| "loss": 0.3378, |
| "step": 8800 |
| }, |
| { |
| "epoch": 22.5, |
| "learning_rate": 1.9820080000000002e-05, |
| "loss": 0.3504, |
| "step": 9000 |
| }, |
| { |
| "epoch": 23.0, |
| "learning_rate": 1.9816080000000004e-05, |
| "loss": 0.3089, |
| "step": 9200 |
| }, |
| { |
| "epoch": 23.5, |
| "learning_rate": 1.98121e-05, |
| "loss": 0.3184, |
| "step": 9400 |
| }, |
| { |
| "epoch": 24.0, |
| "learning_rate": 1.98081e-05, |
| "loss": 0.3183, |
| "step": 9600 |
| }, |
| { |
| "epoch": 24.5, |
| "learning_rate": 1.9804100000000002e-05, |
| "loss": 0.2948, |
| "step": 9800 |
| }, |
| { |
| "epoch": 25.0, |
| "learning_rate": 1.9800100000000004e-05, |
| "loss": 0.3077, |
| "step": 10000 |
| }, |
| { |
| "epoch": 25.0, |
| "eval_f1": 0.7101063829787234, |
| "eval_iou": 0.0038253916427493095, |
| "eval_loss": 0.4940188229084015, |
| "eval_runtime": 19.1335, |
| "eval_samples_per_second": 4.599, |
| "eval_steps_per_second": 2.3, |
| "step": 10000 |
| }, |
| { |
| "epoch": 25.5, |
| "learning_rate": 1.9796100000000002e-05, |
| "loss": 0.2916, |
| "step": 10200 |
| }, |
| { |
| "epoch": 26.0, |
| "learning_rate": 1.97921e-05, |
| "loss": 0.2823, |
| "step": 10400 |
| }, |
| { |
| "epoch": 26.5, |
| "learning_rate": 1.97881e-05, |
| "loss": 0.3004, |
| "step": 10600 |
| }, |
| { |
| "epoch": 27.0, |
| "learning_rate": 1.9784100000000003e-05, |
| "loss": 0.2592, |
| "step": 10800 |
| }, |
| { |
| "epoch": 27.5, |
| "learning_rate": 1.97801e-05, |
| "loss": 0.2724, |
| "step": 11000 |
| }, |
| { |
| "epoch": 28.0, |
| "learning_rate": 1.9776100000000002e-05, |
| "loss": 0.2709, |
| "step": 11200 |
| }, |
| { |
| "epoch": 28.5, |
| "learning_rate": 1.97721e-05, |
| "loss": 0.2633, |
| "step": 11400 |
| }, |
| { |
| "epoch": 29.0, |
| "learning_rate": 1.97681e-05, |
| "loss": 0.2645, |
| "step": 11600 |
| }, |
| { |
| "epoch": 29.5, |
| "learning_rate": 1.9764100000000003e-05, |
| "loss": 0.2649, |
| "step": 11800 |
| }, |
| { |
| "epoch": 30.0, |
| "learning_rate": 1.97601e-05, |
| "loss": 0.2579, |
| "step": 12000 |
| }, |
| { |
| "epoch": 30.0, |
| "eval_f1": 0.764163372859025, |
| "eval_iou": 0.02996533177793026, |
| "eval_loss": 0.4145811200141907, |
| "eval_runtime": 19.0881, |
| "eval_samples_per_second": 4.61, |
| "eval_steps_per_second": 2.305, |
| "step": 12000 |
| }, |
| { |
| "epoch": 30.5, |
| "learning_rate": 1.9756100000000002e-05, |
| "loss": 0.246, |
| "step": 12200 |
| }, |
| { |
| "epoch": 31.0, |
| "learning_rate": 1.97521e-05, |
| "loss": 0.2599, |
| "step": 12400 |
| }, |
| { |
| "epoch": 31.5, |
| "learning_rate": 1.97481e-05, |
| "loss": 0.2519, |
| "step": 12600 |
| }, |
| { |
| "epoch": 32.0, |
| "learning_rate": 1.9744100000000003e-05, |
| "loss": 0.2502, |
| "step": 12800 |
| }, |
| { |
| "epoch": 32.5, |
| "learning_rate": 1.97401e-05, |
| "loss": 0.2589, |
| "step": 13000 |
| }, |
| { |
| "epoch": 33.0, |
| "learning_rate": 1.9736100000000002e-05, |
| "loss": 0.2332, |
| "step": 13200 |
| }, |
| { |
| "epoch": 33.5, |
| "learning_rate": 1.97321e-05, |
| "loss": 0.2436, |
| "step": 13400 |
| }, |
| { |
| "epoch": 34.0, |
| "learning_rate": 1.9728120000000003e-05, |
| "loss": 0.2162, |
| "step": 13600 |
| }, |
| { |
| "epoch": 34.5, |
| "learning_rate": 1.972412e-05, |
| "loss": 0.2532, |
| "step": 13800 |
| }, |
| { |
| "epoch": 35.0, |
| "learning_rate": 1.9720120000000003e-05, |
| "loss": 0.2455, |
| "step": 14000 |
| }, |
| { |
| "epoch": 35.0, |
| "eval_f1": 0.7985257985257985, |
| "eval_iou": 0.027949659153819084, |
| "eval_loss": 0.5306004881858826, |
| "eval_runtime": 18.956, |
| "eval_samples_per_second": 4.642, |
| "eval_steps_per_second": 2.321, |
| "step": 14000 |
| }, |
| { |
| "epoch": 35.5, |
| "learning_rate": 1.971612e-05, |
| "loss": 0.2193, |
| "step": 14200 |
| }, |
| { |
| "epoch": 36.0, |
| "learning_rate": 1.9712120000000002e-05, |
| "loss": 0.2217, |
| "step": 14400 |
| }, |
| { |
| "epoch": 36.5, |
| "learning_rate": 1.9708120000000003e-05, |
| "loss": 0.2165, |
| "step": 14600 |
| }, |
| { |
| "epoch": 37.0, |
| "learning_rate": 1.970412e-05, |
| "loss": 0.2338, |
| "step": 14800 |
| }, |
| { |
| "epoch": 37.5, |
| "learning_rate": 1.9700120000000003e-05, |
| "loss": 0.2215, |
| "step": 15000 |
| }, |
| { |
| "epoch": 38.0, |
| "learning_rate": 1.969612e-05, |
| "loss": 0.2368, |
| "step": 15200 |
| }, |
| { |
| "epoch": 38.5, |
| "learning_rate": 1.9692120000000002e-05, |
| "loss": 0.2328, |
| "step": 15400 |
| }, |
| { |
| "epoch": 39.0, |
| "learning_rate": 1.968812e-05, |
| "loss": 0.225, |
| "step": 15600 |
| }, |
| { |
| "epoch": 39.5, |
| "learning_rate": 1.968416e-05, |
| "loss": 0.2031, |
| "step": 15800 |
| }, |
| { |
| "epoch": 40.0, |
| "learning_rate": 1.9680160000000002e-05, |
| "loss": 0.2355, |
| "step": 16000 |
| }, |
| { |
| "epoch": 40.0, |
| "eval_f1": 0.8187134502923976, |
| "eval_iou": 0.07257882505655289, |
| "eval_loss": 0.40121352672576904, |
| "eval_runtime": 19.0802, |
| "eval_samples_per_second": 4.612, |
| "eval_steps_per_second": 2.306, |
| "step": 16000 |
| }, |
| { |
| "epoch": 40.5, |
| "learning_rate": 1.967616e-05, |
| "loss": 0.2491, |
| "step": 16200 |
| }, |
| { |
| "epoch": 41.0, |
| "learning_rate": 1.9672160000000002e-05, |
| "loss": 0.2006, |
| "step": 16400 |
| }, |
| { |
| "epoch": 41.5, |
| "learning_rate": 1.966816e-05, |
| "loss": 0.218, |
| "step": 16600 |
| }, |
| { |
| "epoch": 42.0, |
| "learning_rate": 1.966416e-05, |
| "loss": 0.2251, |
| "step": 16800 |
| }, |
| { |
| "epoch": 42.5, |
| "learning_rate": 1.9660160000000003e-05, |
| "loss": 0.2039, |
| "step": 17000 |
| }, |
| { |
| "epoch": 43.0, |
| "learning_rate": 1.965616e-05, |
| "loss": 0.2233, |
| "step": 17200 |
| }, |
| { |
| "epoch": 43.5, |
| "learning_rate": 1.9652160000000002e-05, |
| "loss": 0.1836, |
| "step": 17400 |
| }, |
| { |
| "epoch": 44.0, |
| "learning_rate": 1.964816e-05, |
| "loss": 0.2116, |
| "step": 17600 |
| }, |
| { |
| "epoch": 44.5, |
| "learning_rate": 1.964416e-05, |
| "loss": 0.1977, |
| "step": 17800 |
| }, |
| { |
| "epoch": 45.0, |
| "learning_rate": 1.9640160000000003e-05, |
| "loss": 0.2254, |
| "step": 18000 |
| }, |
| { |
| "epoch": 45.0, |
| "eval_f1": 0.8462484624846248, |
| "eval_iou": 0.10463450849056244, |
| "eval_loss": 0.3598695993423462, |
| "eval_runtime": 18.9446, |
| "eval_samples_per_second": 4.645, |
| "eval_steps_per_second": 2.323, |
| "step": 18000 |
| }, |
| { |
| "epoch": 45.5, |
| "learning_rate": 1.963616e-05, |
| "loss": 0.2041, |
| "step": 18200 |
| }, |
| { |
| "epoch": 46.0, |
| "learning_rate": 1.9632160000000002e-05, |
| "loss": 0.2019, |
| "step": 18400 |
| }, |
| { |
| "epoch": 46.5, |
| "learning_rate": 1.962816e-05, |
| "loss": 0.1946, |
| "step": 18600 |
| }, |
| { |
| "epoch": 47.0, |
| "learning_rate": 1.9624180000000003e-05, |
| "loss": 0.225, |
| "step": 18800 |
| }, |
| { |
| "epoch": 47.5, |
| "learning_rate": 1.962018e-05, |
| "loss": 0.1941, |
| "step": 19000 |
| }, |
| { |
| "epoch": 48.0, |
| "learning_rate": 1.9616180000000002e-05, |
| "loss": 0.2288, |
| "step": 19200 |
| }, |
| { |
| "epoch": 48.5, |
| "learning_rate": 1.961218e-05, |
| "loss": 0.1866, |
| "step": 19400 |
| }, |
| { |
| "epoch": 49.0, |
| "learning_rate": 1.9608180000000002e-05, |
| "loss": 0.1843, |
| "step": 19600 |
| }, |
| { |
| "epoch": 49.5, |
| "learning_rate": 1.9604180000000003e-05, |
| "loss": 0.2144, |
| "step": 19800 |
| }, |
| { |
| "epoch": 50.0, |
| "learning_rate": 1.960018e-05, |
| "loss": 0.1965, |
| "step": 20000 |
| }, |
| { |
| "epoch": 50.0, |
| "eval_f1": 0.8717948717948717, |
| "eval_iou": 0.10999788343906403, |
| "eval_loss": 0.32069775462150574, |
| "eval_runtime": 18.9317, |
| "eval_samples_per_second": 4.648, |
| "eval_steps_per_second": 2.324, |
| "step": 20000 |
| }, |
| { |
| "epoch": 50.5, |
| "learning_rate": 1.9596180000000003e-05, |
| "loss": 0.2057, |
| "step": 20200 |
| }, |
| { |
| "epoch": 51.0, |
| "learning_rate": 1.959218e-05, |
| "loss": 0.1912, |
| "step": 20400 |
| }, |
| { |
| "epoch": 51.5, |
| "learning_rate": 1.9588180000000002e-05, |
| "loss": 0.2044, |
| "step": 20600 |
| }, |
| { |
| "epoch": 52.0, |
| "learning_rate": 1.958418e-05, |
| "loss": 0.1809, |
| "step": 20800 |
| }, |
| { |
| "epoch": 52.5, |
| "learning_rate": 1.958018e-05, |
| "loss": 0.2015, |
| "step": 21000 |
| }, |
| { |
| "epoch": 53.0, |
| "learning_rate": 1.95762e-05, |
| "loss": 0.1883, |
| "step": 21200 |
| }, |
| { |
| "epoch": 53.5, |
| "learning_rate": 1.9572200000000002e-05, |
| "loss": 0.1977, |
| "step": 21400 |
| }, |
| { |
| "epoch": 54.0, |
| "learning_rate": 1.95682e-05, |
| "loss": 0.1826, |
| "step": 21600 |
| }, |
| { |
| "epoch": 54.5, |
| "learning_rate": 1.95642e-05, |
| "loss": 0.24, |
| "step": 21800 |
| }, |
| { |
| "epoch": 55.0, |
| "learning_rate": 1.9560200000000003e-05, |
| "loss": 0.1743, |
| "step": 22000 |
| }, |
| { |
| "epoch": 55.0, |
| "eval_f1": 0.8530465949820788, |
| "eval_iou": 0.12947940826416016, |
| "eval_loss": 0.4376421868801117, |
| "eval_runtime": 18.2109, |
| "eval_samples_per_second": 4.832, |
| "eval_steps_per_second": 2.416, |
| "step": 22000 |
| }, |
| { |
| "epoch": 55.5, |
| "learning_rate": 1.95562e-05, |
| "loss": 0.1729, |
| "step": 22200 |
| }, |
| { |
| "epoch": 56.0, |
| "learning_rate": 1.9552200000000002e-05, |
| "loss": 0.1831, |
| "step": 22400 |
| }, |
| { |
| "epoch": 56.5, |
| "learning_rate": 1.95482e-05, |
| "loss": 0.1728, |
| "step": 22600 |
| }, |
| { |
| "epoch": 57.0, |
| "learning_rate": 1.9544200000000002e-05, |
| "loss": 0.1958, |
| "step": 22800 |
| }, |
| { |
| "epoch": 57.5, |
| "learning_rate": 1.9540200000000003e-05, |
| "loss": 0.1872, |
| "step": 23000 |
| }, |
| { |
| "epoch": 58.0, |
| "learning_rate": 1.95362e-05, |
| "loss": 0.1827, |
| "step": 23200 |
| }, |
| { |
| "epoch": 58.5, |
| "learning_rate": 1.9532200000000002e-05, |
| "loss": 0.1985, |
| "step": 23400 |
| }, |
| { |
| "epoch": 59.0, |
| "learning_rate": 1.95282e-05, |
| "loss": 0.1686, |
| "step": 23600 |
| }, |
| { |
| "epoch": 59.5, |
| "learning_rate": 1.9524200000000002e-05, |
| "loss": 0.1748, |
| "step": 23800 |
| }, |
| { |
| "epoch": 60.0, |
| "learning_rate": 1.9520200000000003e-05, |
| "loss": 0.1871, |
| "step": 24000 |
| }, |
| { |
| "epoch": 60.0, |
| "eval_f1": 0.8664259927797834, |
| "eval_iou": 0.10719826072454453, |
| "eval_loss": 0.3321523666381836, |
| "eval_runtime": 19.2285, |
| "eval_samples_per_second": 4.577, |
| "eval_steps_per_second": 2.288, |
| "step": 24000 |
| }, |
| { |
| "epoch": 60.5, |
| "learning_rate": 1.95162e-05, |
| "loss": 0.1907, |
| "step": 24200 |
| }, |
| { |
| "epoch": 61.0, |
| "learning_rate": 1.9512200000000003e-05, |
| "loss": 0.155, |
| "step": 24400 |
| }, |
| { |
| "epoch": 61.5, |
| "learning_rate": 1.95082e-05, |
| "loss": 0.1872, |
| "step": 24600 |
| }, |
| { |
| "epoch": 62.0, |
| "learning_rate": 1.9504200000000002e-05, |
| "loss": 0.1597, |
| "step": 24800 |
| }, |
| { |
| "epoch": 62.5, |
| "learning_rate": 1.9500200000000003e-05, |
| "loss": 0.1651, |
| "step": 25000 |
| }, |
| { |
| "epoch": 63.0, |
| "learning_rate": 1.94962e-05, |
| "loss": 0.1776, |
| "step": 25200 |
| }, |
| { |
| "epoch": 63.5, |
| "learning_rate": 1.949222e-05, |
| "loss": 0.1528, |
| "step": 25400 |
| }, |
| { |
| "epoch": 64.0, |
| "learning_rate": 1.9488220000000002e-05, |
| "loss": 0.1743, |
| "step": 25600 |
| }, |
| { |
| "epoch": 64.5, |
| "learning_rate": 1.9484220000000004e-05, |
| "loss": 0.1699, |
| "step": 25800 |
| }, |
| { |
| "epoch": 65.0, |
| "learning_rate": 1.948022e-05, |
| "loss": 0.1584, |
| "step": 26000 |
| }, |
| { |
| "epoch": 65.0, |
| "eval_f1": 0.9173457508731083, |
| "eval_iou": 0.18261756002902985, |
| "eval_loss": 0.22134169936180115, |
| "eval_runtime": 18.9349, |
| "eval_samples_per_second": 4.648, |
| "eval_steps_per_second": 2.324, |
| "step": 26000 |
| }, |
| { |
| "epoch": 65.5, |
| "learning_rate": 1.947622e-05, |
| "loss": 0.1736, |
| "step": 26200 |
| }, |
| { |
| "epoch": 66.0, |
| "learning_rate": 1.947222e-05, |
| "loss": 0.166, |
| "step": 26400 |
| }, |
| { |
| "epoch": 66.5, |
| "learning_rate": 1.9468220000000002e-05, |
| "loss": 0.1621, |
| "step": 26600 |
| }, |
| { |
| "epoch": 67.0, |
| "learning_rate": 1.9464220000000004e-05, |
| "loss": 0.1607, |
| "step": 26800 |
| }, |
| { |
| "epoch": 67.5, |
| "learning_rate": 1.9460220000000002e-05, |
| "loss": 0.1454, |
| "step": 27000 |
| }, |
| { |
| "epoch": 68.0, |
| "learning_rate": 1.945622e-05, |
| "loss": 0.1805, |
| "step": 27200 |
| }, |
| { |
| "epoch": 68.5, |
| "learning_rate": 1.945222e-05, |
| "loss": 0.1585, |
| "step": 27400 |
| }, |
| { |
| "epoch": 69.0, |
| "learning_rate": 1.9448220000000003e-05, |
| "loss": 0.1741, |
| "step": 27600 |
| }, |
| { |
| "epoch": 69.5, |
| "learning_rate": 1.9444240000000002e-05, |
| "loss": 0.1575, |
| "step": 27800 |
| }, |
| { |
| "epoch": 70.0, |
| "learning_rate": 1.944024e-05, |
| "loss": 0.1709, |
| "step": 28000 |
| }, |
| { |
| "epoch": 70.0, |
| "eval_f1": 0.9138495092693566, |
| "eval_iou": 0.18008637428283691, |
| "eval_loss": 0.2461702823638916, |
| "eval_runtime": 18.9419, |
| "eval_samples_per_second": 4.646, |
| "eval_steps_per_second": 2.323, |
| "step": 28000 |
| }, |
| { |
| "epoch": 70.5, |
| "learning_rate": 1.943624e-05, |
| "loss": 0.1498, |
| "step": 28200 |
| }, |
| { |
| "epoch": 71.0, |
| "learning_rate": 1.9432240000000003e-05, |
| "loss": 0.1713, |
| "step": 28400 |
| }, |
| { |
| "epoch": 71.5, |
| "learning_rate": 1.9428240000000004e-05, |
| "loss": 0.1516, |
| "step": 28600 |
| }, |
| { |
| "epoch": 72.0, |
| "learning_rate": 1.9424240000000002e-05, |
| "loss": 0.1856, |
| "step": 28800 |
| }, |
| { |
| "epoch": 72.5, |
| "learning_rate": 1.942024e-05, |
| "loss": 0.15, |
| "step": 29000 |
| }, |
| { |
| "epoch": 73.0, |
| "learning_rate": 1.941624e-05, |
| "loss": 0.1864, |
| "step": 29200 |
| }, |
| { |
| "epoch": 73.5, |
| "learning_rate": 1.941226e-05, |
| "loss": 0.1574, |
| "step": 29400 |
| }, |
| { |
| "epoch": 74.0, |
| "learning_rate": 1.9408260000000002e-05, |
| "loss": 0.1593, |
| "step": 29600 |
| }, |
| { |
| "epoch": 74.5, |
| "learning_rate": 1.940426e-05, |
| "loss": 0.154, |
| "step": 29800 |
| }, |
| { |
| "epoch": 75.0, |
| "learning_rate": 1.9400260000000002e-05, |
| "loss": 0.1563, |
| "step": 30000 |
| }, |
| { |
| "epoch": 75.0, |
| "eval_f1": 0.8998946259220232, |
| "eval_iou": 0.19760873913764954, |
| "eval_loss": 0.2587834298610687, |
| "eval_runtime": 18.9354, |
| "eval_samples_per_second": 4.647, |
| "eval_steps_per_second": 2.324, |
| "step": 30000 |
| }, |
| { |
| "epoch": 75.5, |
| "learning_rate": 1.9396260000000003e-05, |
| "loss": 0.1543, |
| "step": 30200 |
| }, |
| { |
| "epoch": 76.0, |
| "learning_rate": 1.939226e-05, |
| "loss": 0.1539, |
| "step": 30400 |
| }, |
| { |
| "epoch": 76.5, |
| "learning_rate": 1.9388260000000003e-05, |
| "loss": 0.1494, |
| "step": 30600 |
| }, |
| { |
| "epoch": 77.0, |
| "learning_rate": 1.938426e-05, |
| "loss": 0.1586, |
| "step": 30800 |
| }, |
| { |
| "epoch": 77.5, |
| "learning_rate": 1.9380260000000002e-05, |
| "loss": 0.1614, |
| "step": 31000 |
| }, |
| { |
| "epoch": 78.0, |
| "learning_rate": 1.9376260000000003e-05, |
| "loss": 0.1463, |
| "step": 31200 |
| }, |
| { |
| "epoch": 78.5, |
| "learning_rate": 1.937226e-05, |
| "loss": 0.1469, |
| "step": 31400 |
| }, |
| { |
| "epoch": 79.0, |
| "learning_rate": 1.936826e-05, |
| "loss": 0.1389, |
| "step": 31600 |
| }, |
| { |
| "epoch": 79.5, |
| "learning_rate": 1.936426e-05, |
| "loss": 0.1495, |
| "step": 31800 |
| }, |
| { |
| "epoch": 80.0, |
| "learning_rate": 1.9360260000000002e-05, |
| "loss": 0.1548, |
| "step": 32000 |
| }, |
| { |
| "epoch": 80.0, |
| "eval_f1": 0.8630806845965769, |
| "eval_iou": 0.12309271097183228, |
| "eval_loss": 0.38629788160324097, |
| "eval_runtime": 18.8459, |
| "eval_samples_per_second": 4.669, |
| "eval_steps_per_second": 2.335, |
| "step": 32000 |
| }, |
| { |
| "epoch": 80.5, |
| "learning_rate": 1.935628e-05, |
| "loss": 0.1598, |
| "step": 32200 |
| }, |
| { |
| "epoch": 81.0, |
| "learning_rate": 1.935228e-05, |
| "loss": 0.1452, |
| "step": 32400 |
| }, |
| { |
| "epoch": 81.5, |
| "learning_rate": 1.934828e-05, |
| "loss": 0.1449, |
| "step": 32600 |
| }, |
| { |
| "epoch": 82.0, |
| "learning_rate": 1.9344280000000002e-05, |
| "loss": 0.1464, |
| "step": 32800 |
| }, |
| { |
| "epoch": 82.5, |
| "learning_rate": 1.9340280000000004e-05, |
| "loss": 0.1384, |
| "step": 33000 |
| }, |
| { |
| "epoch": 83.0, |
| "learning_rate": 1.9336280000000002e-05, |
| "loss": 0.1452, |
| "step": 33200 |
| }, |
| { |
| "epoch": 83.5, |
| "learning_rate": 1.933228e-05, |
| "loss": 0.1541, |
| "step": 33400 |
| }, |
| { |
| "epoch": 84.0, |
| "learning_rate": 1.932828e-05, |
| "loss": 0.1417, |
| "step": 33600 |
| }, |
| { |
| "epoch": 84.5, |
| "learning_rate": 1.9324280000000003e-05, |
| "loss": 0.1481, |
| "step": 33800 |
| }, |
| { |
| "epoch": 85.0, |
| "learning_rate": 1.9320280000000004e-05, |
| "loss": 0.1668, |
| "step": 34000 |
| }, |
| { |
| "epoch": 85.0, |
| "eval_f1": 0.871264367816092, |
| "eval_iou": 0.10027875751256943, |
| "eval_loss": 0.3529146611690521, |
| "eval_runtime": 19.0031, |
| "eval_samples_per_second": 4.631, |
| "eval_steps_per_second": 2.315, |
| "step": 34000 |
| }, |
| { |
| "epoch": 85.5, |
| "learning_rate": 1.9316280000000002e-05, |
| "loss": 0.1261, |
| "step": 34200 |
| }, |
| { |
| "epoch": 86.0, |
| "learning_rate": 1.93123e-05, |
| "loss": 0.1444, |
| "step": 34400 |
| }, |
| { |
| "epoch": 86.5, |
| "learning_rate": 1.9308300000000003e-05, |
| "loss": 0.1288, |
| "step": 34600 |
| }, |
| { |
| "epoch": 87.0, |
| "learning_rate": 1.93043e-05, |
| "loss": 0.1622, |
| "step": 34800 |
| }, |
| { |
| "epoch": 87.5, |
| "learning_rate": 1.9300300000000002e-05, |
| "loss": 0.1375, |
| "step": 35000 |
| }, |
| { |
| "epoch": 88.0, |
| "learning_rate": 1.92963e-05, |
| "loss": 0.1439, |
| "step": 35200 |
| }, |
| { |
| "epoch": 88.5, |
| "learning_rate": 1.92923e-05, |
| "loss": 0.1307, |
| "step": 35400 |
| }, |
| { |
| "epoch": 89.0, |
| "learning_rate": 1.9288300000000003e-05, |
| "loss": 0.1525, |
| "step": 35600 |
| }, |
| { |
| "epoch": 89.5, |
| "learning_rate": 1.92843e-05, |
| "loss": 0.1339, |
| "step": 35800 |
| }, |
| { |
| "epoch": 90.0, |
| "learning_rate": 1.92803e-05, |
| "loss": 0.146, |
| "step": 36000 |
| }, |
| { |
| "epoch": 90.0, |
| "eval_f1": 0.884742041712404, |
| "eval_iou": 0.2148662656545639, |
| "eval_loss": 0.32078301906585693, |
| "eval_runtime": 19.1462, |
| "eval_samples_per_second": 4.596, |
| "eval_steps_per_second": 2.298, |
| "step": 36000 |
| }, |
| { |
| "epoch": 90.5, |
| "learning_rate": 1.92763e-05, |
| "loss": 0.1432, |
| "step": 36200 |
| }, |
| { |
| "epoch": 91.0, |
| "learning_rate": 1.9272300000000002e-05, |
| "loss": 0.1573, |
| "step": 36400 |
| }, |
| { |
| "epoch": 91.5, |
| "learning_rate": 1.9268300000000003e-05, |
| "loss": 0.1268, |
| "step": 36600 |
| }, |
| { |
| "epoch": 92.0, |
| "learning_rate": 1.92643e-05, |
| "loss": 0.1407, |
| "step": 36800 |
| }, |
| { |
| "epoch": 92.5, |
| "learning_rate": 1.92603e-05, |
| "loss": 0.1265, |
| "step": 37000 |
| }, |
| { |
| "epoch": 93.0, |
| "learning_rate": 1.92563e-05, |
| "loss": 0.1371, |
| "step": 37200 |
| }, |
| { |
| "epoch": 93.5, |
| "learning_rate": 1.9252300000000002e-05, |
| "loss": 0.129, |
| "step": 37400 |
| }, |
| { |
| "epoch": 94.0, |
| "learning_rate": 1.924832e-05, |
| "loss": 0.1361, |
| "step": 37600 |
| }, |
| { |
| "epoch": 94.5, |
| "learning_rate": 1.924432e-05, |
| "loss": 0.129, |
| "step": 37800 |
| }, |
| { |
| "epoch": 95.0, |
| "learning_rate": 1.924032e-05, |
| "loss": 0.1476, |
| "step": 38000 |
| }, |
| { |
| "epoch": 95.0, |
| "eval_f1": 0.8731082654249126, |
| "eval_iou": 0.20559382438659668, |
| "eval_loss": 0.38162490725517273, |
| "eval_runtime": 19.0093, |
| "eval_samples_per_second": 4.629, |
| "eval_steps_per_second": 2.315, |
| "step": 38000 |
| }, |
| { |
| "epoch": 95.5, |
| "learning_rate": 1.9236320000000002e-05, |
| "loss": 0.1528, |
| "step": 38200 |
| }, |
| { |
| "epoch": 96.0, |
| "learning_rate": 1.9232320000000004e-05, |
| "loss": 0.1232, |
| "step": 38400 |
| }, |
| { |
| "epoch": 96.5, |
| "learning_rate": 1.922832e-05, |
| "loss": 0.1418, |
| "step": 38600 |
| }, |
| { |
| "epoch": 97.0, |
| "learning_rate": 1.922432e-05, |
| "loss": 0.1339, |
| "step": 38800 |
| }, |
| { |
| "epoch": 97.5, |
| "learning_rate": 1.922032e-05, |
| "loss": 0.1298, |
| "step": 39000 |
| }, |
| { |
| "epoch": 98.0, |
| "learning_rate": 1.9216320000000002e-05, |
| "loss": 0.1186, |
| "step": 39200 |
| }, |
| { |
| "epoch": 98.5, |
| "learning_rate": 1.921232e-05, |
| "loss": 0.1332, |
| "step": 39400 |
| }, |
| { |
| "epoch": 99.0, |
| "learning_rate": 1.9208320000000002e-05, |
| "loss": 0.1273, |
| "step": 39600 |
| }, |
| { |
| "epoch": 99.5, |
| "learning_rate": 1.920432e-05, |
| "loss": 0.1329, |
| "step": 39800 |
| }, |
| { |
| "epoch": 100.0, |
| "learning_rate": 1.920032e-05, |
| "loss": 0.1328, |
| "step": 40000 |
| }, |
| { |
| "epoch": 100.0, |
| "eval_f1": 0.8807339449541284, |
| "eval_iou": 0.20764991641044617, |
| "eval_loss": 0.3659871816635132, |
| "eval_runtime": 19.1672, |
| "eval_samples_per_second": 4.591, |
| "eval_steps_per_second": 2.296, |
| "step": 40000 |
| }, |
| { |
| "epoch": 100.5, |
| "learning_rate": 1.9196320000000003e-05, |
| "loss": 0.134, |
| "step": 40200 |
| }, |
| { |
| "epoch": 101.0, |
| "learning_rate": 1.919232e-05, |
| "loss": 0.1266, |
| "step": 40400 |
| }, |
| { |
| "epoch": 101.5, |
| "learning_rate": 1.9188320000000002e-05, |
| "loss": 0.1509, |
| "step": 40600 |
| }, |
| { |
| "epoch": 102.0, |
| "learning_rate": 1.918432e-05, |
| "loss": 0.1185, |
| "step": 40800 |
| }, |
| { |
| "epoch": 102.5, |
| "learning_rate": 1.918032e-05, |
| "loss": 0.1143, |
| "step": 41000 |
| }, |
| { |
| "epoch": 103.0, |
| "learning_rate": 1.917634e-05, |
| "loss": 0.143, |
| "step": 41200 |
| }, |
| { |
| "epoch": 103.5, |
| "learning_rate": 1.9172340000000002e-05, |
| "loss": 0.1251, |
| "step": 41400 |
| }, |
| { |
| "epoch": 104.0, |
| "learning_rate": 1.916834e-05, |
| "loss": 0.1316, |
| "step": 41600 |
| }, |
| { |
| "epoch": 104.5, |
| "learning_rate": 1.916434e-05, |
| "loss": 0.1171, |
| "step": 41800 |
| }, |
| { |
| "epoch": 105.0, |
| "learning_rate": 1.9160340000000003e-05, |
| "loss": 0.1388, |
| "step": 42000 |
| }, |
| { |
| "epoch": 105.0, |
| "eval_f1": 0.9178532311062432, |
| "eval_iou": 0.20952488481998444, |
| "eval_loss": 0.23598887026309967, |
| "eval_runtime": 18.6063, |
| "eval_samples_per_second": 4.73, |
| "eval_steps_per_second": 2.365, |
| "step": 42000 |
| }, |
| { |
| "epoch": 105.5, |
| "learning_rate": 1.915634e-05, |
| "loss": 0.1212, |
| "step": 42200 |
| }, |
| { |
| "epoch": 106.0, |
| "learning_rate": 1.9152340000000002e-05, |
| "loss": 0.1094, |
| "step": 42400 |
| }, |
| { |
| "epoch": 106.5, |
| "learning_rate": 1.914834e-05, |
| "loss": 0.134, |
| "step": 42600 |
| }, |
| { |
| "epoch": 107.0, |
| "learning_rate": 1.9144340000000002e-05, |
| "loss": 0.1276, |
| "step": 42800 |
| }, |
| { |
| "epoch": 107.5, |
| "learning_rate": 1.9140340000000003e-05, |
| "loss": 0.1302, |
| "step": 43000 |
| }, |
| { |
| "epoch": 108.0, |
| "learning_rate": 1.913634e-05, |
| "loss": 0.1139, |
| "step": 43200 |
| }, |
| { |
| "epoch": 108.5, |
| "learning_rate": 1.9132340000000003e-05, |
| "loss": 0.1375, |
| "step": 43400 |
| }, |
| { |
| "epoch": 109.0, |
| "learning_rate": 1.912834e-05, |
| "loss": 0.1215, |
| "step": 43600 |
| }, |
| { |
| "epoch": 109.5, |
| "learning_rate": 1.9124340000000002e-05, |
| "loss": 0.123, |
| "step": 43800 |
| }, |
| { |
| "epoch": 110.0, |
| "learning_rate": 1.9120340000000003e-05, |
| "loss": 0.126, |
| "step": 44000 |
| }, |
| { |
| "epoch": 110.0, |
| "eval_f1": 0.9082969432314411, |
| "eval_iou": 0.23878733813762665, |
| "eval_loss": 0.2843697965145111, |
| "eval_runtime": 19.0545, |
| "eval_samples_per_second": 4.618, |
| "eval_steps_per_second": 2.309, |
| "step": 44000 |
| }, |
| { |
| "epoch": 110.5, |
| "learning_rate": 1.911634e-05, |
| "loss": 0.1018, |
| "step": 44200 |
| }, |
| { |
| "epoch": 111.0, |
| "learning_rate": 1.9112340000000003e-05, |
| "loss": 0.1342, |
| "step": 44400 |
| }, |
| { |
| "epoch": 111.5, |
| "learning_rate": 1.910834e-05, |
| "loss": 0.126, |
| "step": 44600 |
| }, |
| { |
| "epoch": 112.0, |
| "learning_rate": 1.9104340000000002e-05, |
| "loss": 0.1007, |
| "step": 44800 |
| }, |
| { |
| "epoch": 112.5, |
| "learning_rate": 1.910034e-05, |
| "loss": 0.1092, |
| "step": 45000 |
| }, |
| { |
| "epoch": 113.0, |
| "learning_rate": 1.909634e-05, |
| "loss": 0.1208, |
| "step": 45200 |
| }, |
| { |
| "epoch": 113.5, |
| "learning_rate": 1.909236e-05, |
| "loss": 0.1191, |
| "step": 45400 |
| }, |
| { |
| "epoch": 114.0, |
| "learning_rate": 1.9088360000000002e-05, |
| "loss": 0.1246, |
| "step": 45600 |
| }, |
| { |
| "epoch": 114.5, |
| "learning_rate": 1.908436e-05, |
| "loss": 0.1057, |
| "step": 45800 |
| }, |
| { |
| "epoch": 115.0, |
| "learning_rate": 1.9080360000000002e-05, |
| "loss": 0.1358, |
| "step": 46000 |
| }, |
| { |
| "epoch": 115.0, |
| "eval_f1": 0.9072625698324023, |
| "eval_iou": 0.22642718255519867, |
| "eval_loss": 0.28193461894989014, |
| "eval_runtime": 18.847, |
| "eval_samples_per_second": 4.669, |
| "eval_steps_per_second": 2.335, |
| "step": 46000 |
| }, |
| { |
| "epoch": 115.5, |
| "learning_rate": 1.9076360000000003e-05, |
| "loss": 0.1357, |
| "step": 46200 |
| }, |
| { |
| "epoch": 116.0, |
| "learning_rate": 1.907236e-05, |
| "loss": 0.1232, |
| "step": 46400 |
| }, |
| { |
| "epoch": 116.5, |
| "learning_rate": 1.906838e-05, |
| "loss": 0.1083, |
| "step": 46600 |
| }, |
| { |
| "epoch": 117.0, |
| "learning_rate": 1.9064380000000002e-05, |
| "loss": 0.1312, |
| "step": 46800 |
| }, |
| { |
| "epoch": 117.5, |
| "learning_rate": 1.906038e-05, |
| "loss": 0.1182, |
| "step": 47000 |
| }, |
| { |
| "epoch": 118.0, |
| "learning_rate": 1.905638e-05, |
| "loss": 0.1271, |
| "step": 47200 |
| }, |
| { |
| "epoch": 118.5, |
| "learning_rate": 1.9052380000000003e-05, |
| "loss": 0.097, |
| "step": 47400 |
| }, |
| { |
| "epoch": 119.0, |
| "learning_rate": 1.904838e-05, |
| "loss": 0.1272, |
| "step": 47600 |
| }, |
| { |
| "epoch": 119.5, |
| "learning_rate": 1.9044380000000002e-05, |
| "loss": 0.1318, |
| "step": 47800 |
| }, |
| { |
| "epoch": 120.0, |
| "learning_rate": 1.904038e-05, |
| "loss": 0.1092, |
| "step": 48000 |
| }, |
| { |
| "epoch": 120.0, |
| "eval_f1": 0.9151047409040793, |
| "eval_iou": 0.22675490379333496, |
| "eval_loss": 0.3041003346443176, |
| "eval_runtime": 18.9312, |
| "eval_samples_per_second": 4.648, |
| "eval_steps_per_second": 2.324, |
| "step": 48000 |
| }, |
| { |
| "epoch": 120.5, |
| "learning_rate": 1.903638e-05, |
| "loss": 0.1192, |
| "step": 48200 |
| }, |
| { |
| "epoch": 121.0, |
| "learning_rate": 1.9032380000000003e-05, |
| "loss": 0.1073, |
| "step": 48400 |
| }, |
| { |
| "epoch": 121.5, |
| "learning_rate": 1.902838e-05, |
| "loss": 0.1242, |
| "step": 48600 |
| }, |
| { |
| "epoch": 122.0, |
| "learning_rate": 1.9024380000000002e-05, |
| "loss": 0.1205, |
| "step": 48800 |
| }, |
| { |
| "epoch": 122.5, |
| "learning_rate": 1.902038e-05, |
| "loss": 0.1245, |
| "step": 49000 |
| }, |
| { |
| "epoch": 123.0, |
| "learning_rate": 1.9016380000000002e-05, |
| "loss": 0.0985, |
| "step": 49200 |
| }, |
| { |
| "epoch": 123.5, |
| "learning_rate": 1.90124e-05, |
| "loss": 0.111, |
| "step": 49400 |
| }, |
| { |
| "epoch": 124.0, |
| "learning_rate": 1.9008400000000003e-05, |
| "loss": 0.127, |
| "step": 49600 |
| }, |
| { |
| "epoch": 124.5, |
| "learning_rate": 1.90044e-05, |
| "loss": 0.1017, |
| "step": 49800 |
| }, |
| { |
| "epoch": 125.0, |
| "learning_rate": 1.9000400000000002e-05, |
| "loss": 0.1177, |
| "step": 50000 |
| }, |
| { |
| "epoch": 125.0, |
| "eval_f1": 0.922566371681416, |
| "eval_iou": 0.24763403832912445, |
| "eval_loss": 0.24548836052417755, |
| "eval_runtime": 18.9504, |
| "eval_samples_per_second": 4.644, |
| "eval_steps_per_second": 2.322, |
| "step": 50000 |
| }, |
| { |
| "epoch": 125.5, |
| "learning_rate": 1.89964e-05, |
| "loss": 0.1252, |
| "step": 50200 |
| }, |
| { |
| "epoch": 126.0, |
| "learning_rate": 1.89924e-05, |
| "loss": 0.1246, |
| "step": 50400 |
| }, |
| { |
| "epoch": 126.5, |
| "learning_rate": 1.8988400000000003e-05, |
| "loss": 0.1102, |
| "step": 50600 |
| }, |
| { |
| "epoch": 127.0, |
| "learning_rate": 1.89844e-05, |
| "loss": 0.1221, |
| "step": 50800 |
| }, |
| { |
| "epoch": 127.5, |
| "learning_rate": 1.8980400000000002e-05, |
| "loss": 0.1254, |
| "step": 51000 |
| }, |
| { |
| "epoch": 128.0, |
| "learning_rate": 1.89764e-05, |
| "loss": 0.1129, |
| "step": 51200 |
| }, |
| { |
| "epoch": 128.5, |
| "learning_rate": 1.89724e-05, |
| "loss": 0.095, |
| "step": 51400 |
| }, |
| { |
| "epoch": 129.0, |
| "learning_rate": 1.8968400000000003e-05, |
| "loss": 0.1169, |
| "step": 51600 |
| }, |
| { |
| "epoch": 129.5, |
| "learning_rate": 1.89644e-05, |
| "loss": 0.1248, |
| "step": 51800 |
| }, |
| { |
| "epoch": 130.0, |
| "learning_rate": 1.8960400000000002e-05, |
| "loss": 0.1235, |
| "step": 52000 |
| }, |
| { |
| "epoch": 130.0, |
| "eval_f1": 0.8990228013029316, |
| "eval_iou": 0.2005741149187088, |
| "eval_loss": 0.3107660114765167, |
| "eval_runtime": 19.011, |
| "eval_samples_per_second": 4.629, |
| "eval_steps_per_second": 2.314, |
| "step": 52000 |
| }, |
| { |
| "epoch": 130.5, |
| "learning_rate": 1.89564e-05, |
| "loss": 0.0976, |
| "step": 52200 |
| }, |
| { |
| "epoch": 131.0, |
| "learning_rate": 1.8952420000000003e-05, |
| "loss": 0.1074, |
| "step": 52400 |
| }, |
| { |
| "epoch": 131.5, |
| "learning_rate": 1.894842e-05, |
| "loss": 0.1212, |
| "step": 52600 |
| }, |
| { |
| "epoch": 132.0, |
| "learning_rate": 1.8944420000000003e-05, |
| "loss": 0.1065, |
| "step": 52800 |
| }, |
| { |
| "epoch": 132.5, |
| "learning_rate": 1.894042e-05, |
| "loss": 0.0798, |
| "step": 53000 |
| }, |
| { |
| "epoch": 133.0, |
| "learning_rate": 1.8936420000000002e-05, |
| "loss": 0.1238, |
| "step": 53200 |
| }, |
| { |
| "epoch": 133.5, |
| "learning_rate": 1.8932420000000003e-05, |
| "loss": 0.1014, |
| "step": 53400 |
| }, |
| { |
| "epoch": 134.0, |
| "learning_rate": 1.892842e-05, |
| "loss": 0.1036, |
| "step": 53600 |
| }, |
| { |
| "epoch": 134.5, |
| "learning_rate": 1.892444e-05, |
| "loss": 0.1268, |
| "step": 53800 |
| }, |
| { |
| "epoch": 135.0, |
| "learning_rate": 1.8920440000000002e-05, |
| "loss": 0.0916, |
| "step": 54000 |
| }, |
| { |
| "epoch": 135.0, |
| "eval_f1": 0.9165763813651139, |
| "eval_iou": 0.21953877806663513, |
| "eval_loss": 0.29882878065109253, |
| "eval_runtime": 18.8649, |
| "eval_samples_per_second": 4.665, |
| "eval_steps_per_second": 2.332, |
| "step": 54000 |
| }, |
| { |
| "epoch": 135.5, |
| "learning_rate": 1.8916440000000003e-05, |
| "loss": 0.1169, |
| "step": 54200 |
| }, |
| { |
| "epoch": 136.0, |
| "learning_rate": 1.891244e-05, |
| "loss": 0.1018, |
| "step": 54400 |
| }, |
| { |
| "epoch": 136.5, |
| "learning_rate": 1.8908440000000003e-05, |
| "loss": 0.1122, |
| "step": 54600 |
| }, |
| { |
| "epoch": 137.0, |
| "learning_rate": 1.890444e-05, |
| "loss": 0.1095, |
| "step": 54800 |
| }, |
| { |
| "epoch": 137.5, |
| "learning_rate": 1.8900440000000002e-05, |
| "loss": 0.1081, |
| "step": 55000 |
| }, |
| { |
| "epoch": 138.0, |
| "learning_rate": 1.8896440000000004e-05, |
| "loss": 0.1159, |
| "step": 55200 |
| }, |
| { |
| "epoch": 138.5, |
| "learning_rate": 1.889244e-05, |
| "loss": 0.104, |
| "step": 55400 |
| }, |
| { |
| "epoch": 139.0, |
| "learning_rate": 1.888844e-05, |
| "loss": 0.0999, |
| "step": 55600 |
| }, |
| { |
| "epoch": 139.5, |
| "learning_rate": 1.888444e-05, |
| "loss": 0.1135, |
| "step": 55800 |
| }, |
| { |
| "epoch": 140.0, |
| "learning_rate": 1.8880440000000002e-05, |
| "loss": 0.1033, |
| "step": 56000 |
| }, |
| { |
| "epoch": 140.0, |
| "eval_f1": 0.9126637554585152, |
| "eval_iou": 0.24201731383800507, |
| "eval_loss": 0.2935643494129181, |
| "eval_runtime": 18.8975, |
| "eval_samples_per_second": 4.657, |
| "eval_steps_per_second": 2.328, |
| "step": 56000 |
| }, |
| { |
| "epoch": 140.5, |
| "learning_rate": 1.8876440000000004e-05, |
| "loss": 0.1103, |
| "step": 56200 |
| }, |
| { |
| "epoch": 141.0, |
| "learning_rate": 1.8872440000000002e-05, |
| "loss": 0.1115, |
| "step": 56400 |
| }, |
| { |
| "epoch": 141.5, |
| "learning_rate": 1.886844e-05, |
| "loss": 0.0898, |
| "step": 56600 |
| }, |
| { |
| "epoch": 142.0, |
| "learning_rate": 1.886444e-05, |
| "loss": 0.1265, |
| "step": 56800 |
| }, |
| { |
| "epoch": 142.5, |
| "learning_rate": 1.8860440000000003e-05, |
| "loss": 0.0979, |
| "step": 57000 |
| }, |
| { |
| "epoch": 143.0, |
| "learning_rate": 1.8856440000000004e-05, |
| "loss": 0.0972, |
| "step": 57200 |
| }, |
| { |
| "epoch": 143.5, |
| "learning_rate": 1.8852440000000002e-05, |
| "loss": 0.0925, |
| "step": 57400 |
| }, |
| { |
| "epoch": 144.0, |
| "learning_rate": 1.884844e-05, |
| "loss": 0.1098, |
| "step": 57600 |
| }, |
| { |
| "epoch": 144.5, |
| "learning_rate": 1.884444e-05, |
| "loss": 0.0982, |
| "step": 57800 |
| }, |
| { |
| "epoch": 145.0, |
| "learning_rate": 1.8840440000000003e-05, |
| "loss": 0.1022, |
| "step": 58000 |
| }, |
| { |
| "epoch": 145.0, |
| "eval_f1": 0.889132821075741, |
| "eval_iou": 0.2561708986759186, |
| "eval_loss": 0.35731014609336853, |
| "eval_runtime": 18.9226, |
| "eval_samples_per_second": 4.651, |
| "eval_steps_per_second": 2.325, |
| "step": 58000 |
| }, |
| { |
| "epoch": 145.5, |
| "learning_rate": 1.883644e-05, |
| "loss": 0.1032, |
| "step": 58200 |
| }, |
| { |
| "epoch": 146.0, |
| "learning_rate": 1.8832440000000002e-05, |
| "loss": 0.1158, |
| "step": 58400 |
| }, |
| { |
| "epoch": 146.5, |
| "learning_rate": 1.882844e-05, |
| "loss": 0.1187, |
| "step": 58600 |
| }, |
| { |
| "epoch": 147.0, |
| "learning_rate": 1.8824460000000003e-05, |
| "loss": 0.1133, |
| "step": 58800 |
| }, |
| { |
| "epoch": 147.5, |
| "learning_rate": 1.882046e-05, |
| "loss": 0.1113, |
| "step": 59000 |
| }, |
| { |
| "epoch": 148.0, |
| "learning_rate": 1.8816460000000002e-05, |
| "loss": 0.0997, |
| "step": 59200 |
| }, |
| { |
| "epoch": 148.5, |
| "learning_rate": 1.881246e-05, |
| "loss": 0.1192, |
| "step": 59400 |
| }, |
| { |
| "epoch": 149.0, |
| "learning_rate": 1.8808460000000002e-05, |
| "loss": 0.1084, |
| "step": 59600 |
| }, |
| { |
| "epoch": 149.5, |
| "learning_rate": 1.8804460000000003e-05, |
| "loss": 0.0895, |
| "step": 59800 |
| }, |
| { |
| "epoch": 150.0, |
| "learning_rate": 1.880046e-05, |
| "loss": 0.1348, |
| "step": 60000 |
| }, |
| { |
| "epoch": 150.0, |
| "eval_f1": 0.9031505250875146, |
| "eval_iou": 0.23726467788219452, |
| "eval_loss": 0.27234533429145813, |
| "eval_runtime": 19.0861, |
| "eval_samples_per_second": 4.611, |
| "eval_steps_per_second": 2.305, |
| "step": 60000 |
| }, |
| { |
| "epoch": 150.5, |
| "learning_rate": 1.879646e-05, |
| "loss": 0.109, |
| "step": 60200 |
| }, |
| { |
| "epoch": 151.0, |
| "learning_rate": 1.879246e-05, |
| "loss": 0.0921, |
| "step": 60400 |
| }, |
| { |
| "epoch": 151.5, |
| "learning_rate": 1.8788460000000002e-05, |
| "loss": 0.0989, |
| "step": 60600 |
| }, |
| { |
| "epoch": 152.0, |
| "learning_rate": 1.8784460000000003e-05, |
| "loss": 0.076, |
| "step": 60800 |
| }, |
| { |
| "epoch": 152.5, |
| "learning_rate": 1.878046e-05, |
| "loss": 0.1078, |
| "step": 61000 |
| }, |
| { |
| "epoch": 153.0, |
| "learning_rate": 1.877646e-05, |
| "loss": 0.0913, |
| "step": 61200 |
| }, |
| { |
| "epoch": 153.5, |
| "learning_rate": 1.877246e-05, |
| "loss": 0.0915, |
| "step": 61400 |
| }, |
| { |
| "epoch": 154.0, |
| "learning_rate": 1.8768460000000002e-05, |
| "loss": 0.1114, |
| "step": 61600 |
| }, |
| { |
| "epoch": 154.5, |
| "learning_rate": 1.876448e-05, |
| "loss": 0.109, |
| "step": 61800 |
| }, |
| { |
| "epoch": 155.0, |
| "learning_rate": 1.876048e-05, |
| "loss": 0.1207, |
| "step": 62000 |
| }, |
| { |
| "epoch": 155.0, |
| "eval_f1": 0.9106753812636166, |
| "eval_iou": 0.24303166568279266, |
| "eval_loss": 0.31264910101890564, |
| "eval_runtime": 19.0168, |
| "eval_samples_per_second": 4.627, |
| "eval_steps_per_second": 2.314, |
| "step": 62000 |
| }, |
| { |
| "epoch": 155.5, |
| "learning_rate": 1.875648e-05, |
| "loss": 0.1077, |
| "step": 62200 |
| }, |
| { |
| "epoch": 156.0, |
| "learning_rate": 1.8752480000000002e-05, |
| "loss": 0.0875, |
| "step": 62400 |
| }, |
| { |
| "epoch": 156.5, |
| "learning_rate": 1.8748480000000004e-05, |
| "loss": 0.1216, |
| "step": 62600 |
| }, |
| { |
| "epoch": 157.0, |
| "learning_rate": 1.8744480000000002e-05, |
| "loss": 0.0917, |
| "step": 62800 |
| }, |
| { |
| "epoch": 157.5, |
| "learning_rate": 1.874048e-05, |
| "loss": 0.1165, |
| "step": 63000 |
| }, |
| { |
| "epoch": 158.0, |
| "learning_rate": 1.873648e-05, |
| "loss": 0.0928, |
| "step": 63200 |
| }, |
| { |
| "epoch": 158.5, |
| "learning_rate": 1.8732480000000003e-05, |
| "loss": 0.1054, |
| "step": 63400 |
| }, |
| { |
| "epoch": 159.0, |
| "learning_rate": 1.872848e-05, |
| "loss": 0.0913, |
| "step": 63600 |
| }, |
| { |
| "epoch": 159.5, |
| "learning_rate": 1.8724480000000002e-05, |
| "loss": 0.0919, |
| "step": 63800 |
| }, |
| { |
| "epoch": 160.0, |
| "learning_rate": 1.872048e-05, |
| "loss": 0.0959, |
| "step": 64000 |
| }, |
| { |
| "epoch": 160.0, |
| "eval_f1": 0.9078651685393258, |
| "eval_iou": 0.1830819547176361, |
| "eval_loss": 0.2727409601211548, |
| "eval_runtime": 18.8918, |
| "eval_samples_per_second": 4.658, |
| "eval_steps_per_second": 2.329, |
| "step": 64000 |
| }, |
| { |
| "epoch": 160.5, |
| "learning_rate": 1.871648e-05, |
| "loss": 0.0878, |
| "step": 64200 |
| }, |
| { |
| "epoch": 161.0, |
| "learning_rate": 1.8712480000000003e-05, |
| "loss": 0.1157, |
| "step": 64400 |
| }, |
| { |
| "epoch": 161.5, |
| "learning_rate": 1.870848e-05, |
| "loss": 0.1029, |
| "step": 64600 |
| }, |
| { |
| "epoch": 162.0, |
| "learning_rate": 1.87045e-05, |
| "loss": 0.0964, |
| "step": 64800 |
| }, |
| { |
| "epoch": 162.5, |
| "learning_rate": 1.87005e-05, |
| "loss": 0.0776, |
| "step": 65000 |
| }, |
| { |
| "epoch": 163.0, |
| "learning_rate": 1.8696500000000003e-05, |
| "loss": 0.1115, |
| "step": 65200 |
| }, |
| { |
| "epoch": 163.5, |
| "learning_rate": 1.86925e-05, |
| "loss": 0.0932, |
| "step": 65400 |
| }, |
| { |
| "epoch": 164.0, |
| "learning_rate": 1.86885e-05, |
| "loss": 0.1149, |
| "step": 65600 |
| }, |
| { |
| "epoch": 164.5, |
| "learning_rate": 1.86845e-05, |
| "loss": 0.114, |
| "step": 65800 |
| }, |
| { |
| "epoch": 165.0, |
| "learning_rate": 1.8680500000000002e-05, |
| "loss": 0.0948, |
| "step": 66000 |
| }, |
| { |
| "epoch": 165.0, |
| "eval_f1": 0.918918918918919, |
| "eval_iou": 0.2549566328525543, |
| "eval_loss": 0.2784244418144226, |
| "eval_runtime": 18.9389, |
| "eval_samples_per_second": 4.647, |
| "eval_steps_per_second": 2.323, |
| "step": 66000 |
| }, |
| { |
| "epoch": 165.5, |
| "learning_rate": 1.8676500000000003e-05, |
| "loss": 0.0963, |
| "step": 66200 |
| }, |
| { |
| "epoch": 166.0, |
| "learning_rate": 1.86725e-05, |
| "loss": 0.1062, |
| "step": 66400 |
| }, |
| { |
| "epoch": 166.5, |
| "learning_rate": 1.86685e-05, |
| "loss": 0.0963, |
| "step": 66600 |
| }, |
| { |
| "epoch": 167.0, |
| "learning_rate": 1.86645e-05, |
| "loss": 0.0872, |
| "step": 66800 |
| }, |
| { |
| "epoch": 167.5, |
| "learning_rate": 1.8660500000000002e-05, |
| "loss": 0.0988, |
| "step": 67000 |
| }, |
| { |
| "epoch": 168.0, |
| "learning_rate": 1.8656500000000003e-05, |
| "loss": 0.1222, |
| "step": 67200 |
| }, |
| { |
| "epoch": 168.5, |
| "learning_rate": 1.86525e-05, |
| "loss": 0.0825, |
| "step": 67400 |
| }, |
| { |
| "epoch": 169.0, |
| "learning_rate": 1.86485e-05, |
| "loss": 0.1053, |
| "step": 67600 |
| }, |
| { |
| "epoch": 169.5, |
| "learning_rate": 1.86445e-05, |
| "loss": 0.0906, |
| "step": 67800 |
| }, |
| { |
| "epoch": 170.0, |
| "learning_rate": 1.8640500000000002e-05, |
| "loss": 0.1127, |
| "step": 68000 |
| }, |
| { |
| "epoch": 170.0, |
| "eval_f1": 0.9211711711711712, |
| "eval_iou": 0.2462574690580368, |
| "eval_loss": 0.24512901902198792, |
| "eval_runtime": 19.1469, |
| "eval_samples_per_second": 4.596, |
| "eval_steps_per_second": 2.298, |
| "step": 68000 |
| }, |
| { |
| "epoch": 170.5, |
| "learning_rate": 1.86365e-05, |
| "loss": 0.0894, |
| "step": 68200 |
| }, |
| { |
| "epoch": 171.0, |
| "learning_rate": 1.86325e-05, |
| "loss": 0.118, |
| "step": 68400 |
| }, |
| { |
| "epoch": 171.5, |
| "learning_rate": 1.862852e-05, |
| "loss": 0.1112, |
| "step": 68600 |
| }, |
| { |
| "epoch": 172.0, |
| "learning_rate": 1.8624520000000002e-05, |
| "loss": 0.1083, |
| "step": 68800 |
| }, |
| { |
| "epoch": 172.5, |
| "learning_rate": 1.862052e-05, |
| "loss": 0.0889, |
| "step": 69000 |
| }, |
| { |
| "epoch": 173.0, |
| "learning_rate": 1.8616520000000002e-05, |
| "loss": 0.1103, |
| "step": 69200 |
| }, |
| { |
| "epoch": 173.5, |
| "learning_rate": 1.861252e-05, |
| "loss": 0.1009, |
| "step": 69400 |
| }, |
| { |
| "epoch": 174.0, |
| "learning_rate": 1.860852e-05, |
| "loss": 0.0891, |
| "step": 69600 |
| }, |
| { |
| "epoch": 174.5, |
| "learning_rate": 1.8604520000000003e-05, |
| "loss": 0.0821, |
| "step": 69800 |
| }, |
| { |
| "epoch": 175.0, |
| "learning_rate": 1.860052e-05, |
| "loss": 0.0828, |
| "step": 70000 |
| }, |
| { |
| "epoch": 175.0, |
| "eval_f1": 0.9254727474972191, |
| "eval_iou": 0.29734450578689575, |
| "eval_loss": 0.2683922052383423, |
| "eval_runtime": 19.0238, |
| "eval_samples_per_second": 4.626, |
| "eval_steps_per_second": 2.313, |
| "step": 70000 |
| }, |
| { |
| "epoch": 175.5, |
| "learning_rate": 1.8596520000000002e-05, |
| "loss": 0.0944, |
| "step": 70200 |
| }, |
| { |
| "epoch": 176.0, |
| "learning_rate": 1.859252e-05, |
| "loss": 0.0789, |
| "step": 70400 |
| }, |
| { |
| "epoch": 176.5, |
| "learning_rate": 1.858852e-05, |
| "loss": 0.0918, |
| "step": 70600 |
| }, |
| { |
| "epoch": 177.0, |
| "learning_rate": 1.8584520000000003e-05, |
| "loss": 0.0876, |
| "step": 70800 |
| }, |
| { |
| "epoch": 177.5, |
| "learning_rate": 1.858052e-05, |
| "loss": 0.0945, |
| "step": 71000 |
| }, |
| { |
| "epoch": 178.0, |
| "learning_rate": 1.8576520000000002e-05, |
| "loss": 0.0763, |
| "step": 71200 |
| }, |
| { |
| "epoch": 178.5, |
| "learning_rate": 1.857252e-05, |
| "loss": 0.0963, |
| "step": 71400 |
| }, |
| { |
| "epoch": 179.0, |
| "learning_rate": 1.856852e-05, |
| "loss": 0.0962, |
| "step": 71600 |
| }, |
| { |
| "epoch": 179.5, |
| "learning_rate": 1.8564520000000003e-05, |
| "loss": 0.0873, |
| "step": 71800 |
| }, |
| { |
| "epoch": 180.0, |
| "learning_rate": 1.856052e-05, |
| "loss": 0.0844, |
| "step": 72000 |
| }, |
| { |
| "epoch": 180.0, |
| "eval_f1": 0.9182242990654206, |
| "eval_iou": 0.2642989456653595, |
| "eval_loss": 0.2961244285106659, |
| "eval_runtime": 19.0376, |
| "eval_samples_per_second": 4.622, |
| "eval_steps_per_second": 2.311, |
| "step": 72000 |
| }, |
| { |
| "epoch": 180.5, |
| "learning_rate": 1.8556520000000002e-05, |
| "loss": 0.0809, |
| "step": 72200 |
| }, |
| { |
| "epoch": 181.0, |
| "learning_rate": 1.855252e-05, |
| "loss": 0.0883, |
| "step": 72400 |
| }, |
| { |
| "epoch": 181.5, |
| "learning_rate": 1.8548540000000003e-05, |
| "loss": 0.0938, |
| "step": 72600 |
| }, |
| { |
| "epoch": 182.0, |
| "learning_rate": 1.854454e-05, |
| "loss": 0.1054, |
| "step": 72800 |
| }, |
| { |
| "epoch": 182.5, |
| "learning_rate": 1.8540540000000003e-05, |
| "loss": 0.0935, |
| "step": 73000 |
| }, |
| { |
| "epoch": 183.0, |
| "learning_rate": 1.853654e-05, |
| "loss": 0.0908, |
| "step": 73200 |
| }, |
| { |
| "epoch": 183.5, |
| "learning_rate": 1.8532540000000002e-05, |
| "loss": 0.0787, |
| "step": 73400 |
| }, |
| { |
| "epoch": 184.0, |
| "learning_rate": 1.852854e-05, |
| "loss": 0.0893, |
| "step": 73600 |
| }, |
| { |
| "epoch": 184.5, |
| "learning_rate": 1.852454e-05, |
| "loss": 0.084, |
| "step": 73800 |
| }, |
| { |
| "epoch": 185.0, |
| "learning_rate": 1.8520540000000003e-05, |
| "loss": 0.1035, |
| "step": 74000 |
| }, |
| { |
| "epoch": 185.0, |
| "eval_f1": 0.9074074074074073, |
| "eval_iou": 0.22117753326892853, |
| "eval_loss": 0.303509384393692, |
| "eval_runtime": 18.7177, |
| "eval_samples_per_second": 4.701, |
| "eval_steps_per_second": 2.351, |
| "step": 74000 |
| }, |
| { |
| "epoch": 185.5, |
| "learning_rate": 1.851654e-05, |
| "loss": 0.0815, |
| "step": 74200 |
| }, |
| { |
| "epoch": 186.0, |
| "learning_rate": 1.851256e-05, |
| "loss": 0.0893, |
| "step": 74400 |
| }, |
| { |
| "epoch": 186.5, |
| "learning_rate": 1.850856e-05, |
| "loss": 0.0888, |
| "step": 74600 |
| }, |
| { |
| "epoch": 187.0, |
| "learning_rate": 1.8504560000000003e-05, |
| "loss": 0.0854, |
| "step": 74800 |
| }, |
| { |
| "epoch": 187.5, |
| "learning_rate": 1.850056e-05, |
| "loss": 0.0865, |
| "step": 75000 |
| }, |
| { |
| "epoch": 188.0, |
| "learning_rate": 1.8496560000000002e-05, |
| "loss": 0.0975, |
| "step": 75200 |
| }, |
| { |
| "epoch": 188.5, |
| "learning_rate": 1.849256e-05, |
| "loss": 0.0927, |
| "step": 75400 |
| }, |
| { |
| "epoch": 189.0, |
| "learning_rate": 1.8488560000000002e-05, |
| "loss": 0.0686, |
| "step": 75600 |
| }, |
| { |
| "epoch": 189.5, |
| "learning_rate": 1.8484560000000003e-05, |
| "loss": 0.0787, |
| "step": 75800 |
| }, |
| { |
| "epoch": 190.0, |
| "learning_rate": 1.848056e-05, |
| "loss": 0.0947, |
| "step": 76000 |
| }, |
| { |
| "epoch": 190.0, |
| "eval_f1": 0.9193205944798302, |
| "eval_iou": 0.2750553786754608, |
| "eval_loss": 0.2922684848308563, |
| "eval_runtime": 18.8568, |
| "eval_samples_per_second": 4.667, |
| "eval_steps_per_second": 2.333, |
| "step": 76000 |
| }, |
| { |
| "epoch": 190.5, |
| "learning_rate": 1.8476560000000003e-05, |
| "loss": 0.1019, |
| "step": 76200 |
| }, |
| { |
| "epoch": 191.0, |
| "learning_rate": 1.847256e-05, |
| "loss": 0.082, |
| "step": 76400 |
| }, |
| { |
| "epoch": 191.5, |
| "learning_rate": 1.8468580000000003e-05, |
| "loss": 0.0972, |
| "step": 76600 |
| }, |
| { |
| "epoch": 192.0, |
| "learning_rate": 1.846458e-05, |
| "loss": 0.0925, |
| "step": 76800 |
| }, |
| { |
| "epoch": 192.5, |
| "learning_rate": 1.8460580000000003e-05, |
| "loss": 0.1045, |
| "step": 77000 |
| }, |
| { |
| "epoch": 193.0, |
| "learning_rate": 1.845658e-05, |
| "loss": 0.0878, |
| "step": 77200 |
| }, |
| { |
| "epoch": 193.5, |
| "learning_rate": 1.8452580000000002e-05, |
| "loss": 0.0939, |
| "step": 77400 |
| }, |
| { |
| "epoch": 194.0, |
| "learning_rate": 1.8448580000000004e-05, |
| "loss": 0.073, |
| "step": 77600 |
| }, |
| { |
| "epoch": 194.5, |
| "learning_rate": 1.844458e-05, |
| "loss": 0.084, |
| "step": 77800 |
| }, |
| { |
| "epoch": 195.0, |
| "learning_rate": 1.8440580000000003e-05, |
| "loss": 0.1011, |
| "step": 78000 |
| }, |
| { |
| "epoch": 195.0, |
| "eval_f1": 0.8741573033707865, |
| "eval_iou": 0.277492880821228, |
| "eval_loss": 0.37412694096565247, |
| "eval_runtime": 19.0722, |
| "eval_samples_per_second": 4.614, |
| "eval_steps_per_second": 2.307, |
| "step": 78000 |
| }, |
| { |
| "epoch": 195.5, |
| "learning_rate": 1.843658e-05, |
| "loss": 0.0952, |
| "step": 78200 |
| }, |
| { |
| "epoch": 196.0, |
| "learning_rate": 1.8432580000000002e-05, |
| "loss": 0.0726, |
| "step": 78400 |
| }, |
| { |
| "epoch": 196.5, |
| "learning_rate": 1.8428580000000004e-05, |
| "loss": 0.0907, |
| "step": 78600 |
| }, |
| { |
| "epoch": 197.0, |
| "learning_rate": 1.8424580000000002e-05, |
| "loss": 0.0737, |
| "step": 78800 |
| }, |
| { |
| "epoch": 197.5, |
| "learning_rate": 1.84206e-05, |
| "loss": 0.0997, |
| "step": 79000 |
| }, |
| { |
| "epoch": 198.0, |
| "learning_rate": 1.8416600000000003e-05, |
| "loss": 0.0905, |
| "step": 79200 |
| }, |
| { |
| "epoch": 198.5, |
| "learning_rate": 1.84126e-05, |
| "loss": 0.0972, |
| "step": 79400 |
| }, |
| { |
| "epoch": 199.0, |
| "learning_rate": 1.8408600000000002e-05, |
| "loss": 0.0706, |
| "step": 79600 |
| }, |
| { |
| "epoch": 199.5, |
| "learning_rate": 1.84046e-05, |
| "loss": 0.0966, |
| "step": 79800 |
| }, |
| { |
| "epoch": 200.0, |
| "learning_rate": 1.84006e-05, |
| "loss": 0.084, |
| "step": 80000 |
| }, |
| { |
| "epoch": 200.0, |
| "eval_f1": 0.9243119266055047, |
| "eval_iou": 0.2545383870601654, |
| "eval_loss": 0.2839164733886719, |
| "eval_runtime": 19.3102, |
| "eval_samples_per_second": 4.557, |
| "eval_steps_per_second": 2.279, |
| "step": 80000 |
| }, |
| { |
| "epoch": 200.5, |
| "learning_rate": 1.8396600000000003e-05, |
| "loss": 0.0762, |
| "step": 80200 |
| }, |
| { |
| "epoch": 201.0, |
| "learning_rate": 1.83926e-05, |
| "loss": 0.1015, |
| "step": 80400 |
| }, |
| { |
| "epoch": 201.5, |
| "learning_rate": 1.8388600000000002e-05, |
| "loss": 0.0915, |
| "step": 80600 |
| }, |
| { |
| "epoch": 202.0, |
| "learning_rate": 1.83846e-05, |
| "loss": 0.0804, |
| "step": 80800 |
| }, |
| { |
| "epoch": 202.5, |
| "learning_rate": 1.8380620000000003e-05, |
| "loss": 0.0987, |
| "step": 81000 |
| }, |
| { |
| "epoch": 203.0, |
| "learning_rate": 1.837662e-05, |
| "loss": 0.0806, |
| "step": 81200 |
| }, |
| { |
| "epoch": 203.5, |
| "learning_rate": 1.8372620000000002e-05, |
| "loss": 0.0835, |
| "step": 81400 |
| }, |
| { |
| "epoch": 204.0, |
| "learning_rate": 1.836862e-05, |
| "loss": 0.0827, |
| "step": 81600 |
| }, |
| { |
| "epoch": 204.5, |
| "learning_rate": 1.8364620000000002e-05, |
| "loss": 0.0827, |
| "step": 81800 |
| }, |
| { |
| "epoch": 205.0, |
| "learning_rate": 1.8360620000000003e-05, |
| "loss": 0.1019, |
| "step": 82000 |
| }, |
| { |
| "epoch": 205.0, |
| "eval_f1": 0.9012208657047726, |
| "eval_iou": 0.2551538050174713, |
| "eval_loss": 0.28741028904914856, |
| "eval_runtime": 18.7631, |
| "eval_samples_per_second": 4.69, |
| "eval_steps_per_second": 2.345, |
| "step": 82000 |
| }, |
| { |
| "epoch": 205.5, |
| "learning_rate": 1.8356640000000003e-05, |
| "loss": 0.069, |
| "step": 82200 |
| }, |
| { |
| "epoch": 206.0, |
| "learning_rate": 1.835264e-05, |
| "loss": 0.0606, |
| "step": 82400 |
| }, |
| { |
| "epoch": 206.5, |
| "learning_rate": 1.8348640000000002e-05, |
| "loss": 0.0448, |
| "step": 82600 |
| }, |
| { |
| "epoch": 207.0, |
| "learning_rate": 1.8344640000000003e-05, |
| "loss": 0.0664, |
| "step": 82800 |
| }, |
| { |
| "epoch": 207.5, |
| "learning_rate": 1.8340660000000003e-05, |
| "loss": 0.0664, |
| "step": 83000 |
| }, |
| { |
| "epoch": 208.0, |
| "learning_rate": 1.833666e-05, |
| "loss": 0.0788, |
| "step": 83200 |
| }, |
| { |
| "epoch": 208.5, |
| "learning_rate": 1.8332660000000002e-05, |
| "loss": 0.0692, |
| "step": 83400 |
| }, |
| { |
| "epoch": 209.0, |
| "learning_rate": 1.8328660000000004e-05, |
| "loss": 0.0564, |
| "step": 83600 |
| }, |
| { |
| "epoch": 209.5, |
| "learning_rate": 1.832466e-05, |
| "loss": 0.065, |
| "step": 83800 |
| }, |
| { |
| "epoch": 210.0, |
| "learning_rate": 1.8320660000000003e-05, |
| "loss": 0.0751, |
| "step": 84000 |
| }, |
| { |
| "epoch": 210.0, |
| "eval_f1": 0.914409534127844, |
| "eval_iou": 0.27163100242614746, |
| "eval_loss": 0.32041147351264954, |
| "eval_runtime": 18.9333, |
| "eval_samples_per_second": 4.648, |
| "eval_steps_per_second": 2.324, |
| "step": 84000 |
| }, |
| { |
| "epoch": 210.5, |
| "learning_rate": 1.831666e-05, |
| "loss": 0.0518, |
| "step": 84200 |
| }, |
| { |
| "epoch": 211.0, |
| "learning_rate": 1.8312660000000002e-05, |
| "loss": 0.0658, |
| "step": 84400 |
| }, |
| { |
| "epoch": 211.5, |
| "learning_rate": 1.8308660000000004e-05, |
| "loss": 0.0683, |
| "step": 84600 |
| }, |
| { |
| "epoch": 212.0, |
| "learning_rate": 1.8304660000000002e-05, |
| "loss": 0.0567, |
| "step": 84800 |
| }, |
| { |
| "epoch": 212.5, |
| "learning_rate": 1.830066e-05, |
| "loss": 0.0792, |
| "step": 85000 |
| }, |
| { |
| "epoch": 213.0, |
| "learning_rate": 1.82967e-05, |
| "loss": 0.0491, |
| "step": 85200 |
| }, |
| { |
| "epoch": 213.5, |
| "learning_rate": 1.8292700000000002e-05, |
| "loss": 0.0797, |
| "step": 85400 |
| }, |
| { |
| "epoch": 214.0, |
| "learning_rate": 1.82887e-05, |
| "loss": 0.0435, |
| "step": 85600 |
| }, |
| { |
| "epoch": 214.5, |
| "learning_rate": 1.82847e-05, |
| "loss": 0.0837, |
| "step": 85800 |
| }, |
| { |
| "epoch": 215.0, |
| "learning_rate": 1.8280700000000003e-05, |
| "loss": 0.0571, |
| "step": 86000 |
| }, |
| { |
| "epoch": 215.0, |
| "eval_f1": 0.9182115594329335, |
| "eval_iou": 0.2814200222492218, |
| "eval_loss": 0.32233256101608276, |
| "eval_runtime": 18.8315, |
| "eval_samples_per_second": 4.673, |
| "eval_steps_per_second": 2.337, |
| "step": 86000 |
| }, |
| { |
| "epoch": 215.5, |
| "learning_rate": 1.82767e-05, |
| "loss": 0.0523, |
| "step": 86200 |
| }, |
| { |
| "epoch": 216.0, |
| "learning_rate": 1.8272700000000002e-05, |
| "loss": 0.0655, |
| "step": 86400 |
| }, |
| { |
| "epoch": 216.5, |
| "learning_rate": 1.82687e-05, |
| "loss": 0.0691, |
| "step": 86600 |
| }, |
| { |
| "epoch": 217.0, |
| "learning_rate": 1.82647e-05, |
| "loss": 0.0674, |
| "step": 86800 |
| }, |
| { |
| "epoch": 217.5, |
| "learning_rate": 1.8260700000000003e-05, |
| "loss": 0.0576, |
| "step": 87000 |
| }, |
| { |
| "epoch": 218.0, |
| "learning_rate": 1.82567e-05, |
| "loss": 0.0676, |
| "step": 87200 |
| }, |
| { |
| "epoch": 218.5, |
| "learning_rate": 1.8252700000000002e-05, |
| "loss": 0.0636, |
| "step": 87400 |
| }, |
| { |
| "epoch": 219.0, |
| "learning_rate": 1.82487e-05, |
| "loss": 0.0623, |
| "step": 87600 |
| }, |
| { |
| "epoch": 219.5, |
| "learning_rate": 1.82447e-05, |
| "loss": 0.0468, |
| "step": 87800 |
| }, |
| { |
| "epoch": 220.0, |
| "learning_rate": 1.8240700000000003e-05, |
| "loss": 0.0744, |
| "step": 88000 |
| }, |
| { |
| "epoch": 220.0, |
| "eval_f1": 0.9115341545352743, |
| "eval_iou": 0.28989362716674805, |
| "eval_loss": 0.28828859329223633, |
| "eval_runtime": 19.0992, |
| "eval_samples_per_second": 4.608, |
| "eval_steps_per_second": 2.304, |
| "step": 88000 |
| }, |
| { |
| "epoch": 220.5, |
| "learning_rate": 1.823674e-05, |
| "loss": 0.0698, |
| "step": 88200 |
| }, |
| { |
| "epoch": 221.0, |
| "learning_rate": 1.8232740000000002e-05, |
| "loss": 0.0725, |
| "step": 88400 |
| }, |
| { |
| "epoch": 221.5, |
| "learning_rate": 1.8228740000000003e-05, |
| "loss": 0.0619, |
| "step": 88600 |
| }, |
| { |
| "epoch": 222.0, |
| "learning_rate": 1.822474e-05, |
| "loss": 0.0666, |
| "step": 88800 |
| }, |
| { |
| "epoch": 222.5, |
| "learning_rate": 1.8220740000000003e-05, |
| "loss": 0.0673, |
| "step": 89000 |
| }, |
| { |
| "epoch": 223.0, |
| "learning_rate": 1.821674e-05, |
| "loss": 0.0553, |
| "step": 89200 |
| }, |
| { |
| "epoch": 223.5, |
| "learning_rate": 1.8212740000000002e-05, |
| "loss": 0.0525, |
| "step": 89400 |
| }, |
| { |
| "epoch": 224.0, |
| "learning_rate": 1.8208740000000003e-05, |
| "loss": 0.0613, |
| "step": 89600 |
| }, |
| { |
| "epoch": 224.5, |
| "learning_rate": 1.820474e-05, |
| "loss": 0.049, |
| "step": 89800 |
| }, |
| { |
| "epoch": 225.0, |
| "learning_rate": 1.8200740000000003e-05, |
| "loss": 0.0669, |
| "step": 90000 |
| }, |
| { |
| "epoch": 225.0, |
| "eval_f1": 0.9017857142857142, |
| "eval_iou": 0.26697784662246704, |
| "eval_loss": 0.3587724566459656, |
| "eval_runtime": 19.0101, |
| "eval_samples_per_second": 4.629, |
| "eval_steps_per_second": 2.315, |
| "step": 90000 |
| }, |
| { |
| "epoch": 225.5, |
| "learning_rate": 1.819674e-05, |
| "loss": 0.0535, |
| "step": 90200 |
| }, |
| { |
| "epoch": 226.0, |
| "learning_rate": 1.8192740000000002e-05, |
| "loss": 0.0468, |
| "step": 90400 |
| }, |
| { |
| "epoch": 226.5, |
| "learning_rate": 1.8188740000000004e-05, |
| "loss": 0.07, |
| "step": 90600 |
| }, |
| { |
| "epoch": 227.0, |
| "learning_rate": 1.818474e-05, |
| "loss": 0.0577, |
| "step": 90800 |
| }, |
| { |
| "epoch": 227.5, |
| "learning_rate": 1.818074e-05, |
| "loss": 0.0477, |
| "step": 91000 |
| }, |
| { |
| "epoch": 228.0, |
| "learning_rate": 1.817674e-05, |
| "loss": 0.0605, |
| "step": 91200 |
| }, |
| { |
| "epoch": 228.5, |
| "learning_rate": 1.8172740000000002e-05, |
| "loss": 0.0525, |
| "step": 91400 |
| }, |
| { |
| "epoch": 229.0, |
| "learning_rate": 1.8168740000000004e-05, |
| "loss": 0.0715, |
| "step": 91600 |
| }, |
| { |
| "epoch": 229.5, |
| "learning_rate": 1.8164740000000002e-05, |
| "loss": 0.0596, |
| "step": 91800 |
| }, |
| { |
| "epoch": 230.0, |
| "learning_rate": 1.816074e-05, |
| "loss": 0.0629, |
| "step": 92000 |
| }, |
| { |
| "epoch": 230.0, |
| "eval_f1": 0.8628370457209847, |
| "eval_iou": 0.26327431201934814, |
| "eval_loss": 0.48599785566329956, |
| "eval_runtime": 18.8693, |
| "eval_samples_per_second": 4.664, |
| "eval_steps_per_second": 2.332, |
| "step": 92000 |
| }, |
| { |
| "epoch": 230.5, |
| "learning_rate": 1.815674e-05, |
| "loss": 0.048, |
| "step": 92200 |
| }, |
| { |
| "epoch": 231.0, |
| "learning_rate": 1.8152740000000003e-05, |
| "loss": 0.0631, |
| "step": 92400 |
| }, |
| { |
| "epoch": 231.5, |
| "learning_rate": 1.8148740000000004e-05, |
| "loss": 0.0537, |
| "step": 92600 |
| }, |
| { |
| "epoch": 232.0, |
| "learning_rate": 1.8144740000000002e-05, |
| "loss": 0.0555, |
| "step": 92800 |
| }, |
| { |
| "epoch": 232.5, |
| "learning_rate": 1.814074e-05, |
| "loss": 0.0719, |
| "step": 93000 |
| }, |
| { |
| "epoch": 233.0, |
| "learning_rate": 1.8136760000000003e-05, |
| "loss": 0.0547, |
| "step": 93200 |
| }, |
| { |
| "epoch": 233.5, |
| "learning_rate": 1.8132760000000004e-05, |
| "loss": 0.054, |
| "step": 93400 |
| }, |
| { |
| "epoch": 234.0, |
| "learning_rate": 1.8128760000000002e-05, |
| "loss": 0.0558, |
| "step": 93600 |
| }, |
| { |
| "epoch": 234.5, |
| "learning_rate": 1.812476e-05, |
| "loss": 0.0577, |
| "step": 93800 |
| }, |
| { |
| "epoch": 235.0, |
| "learning_rate": 1.812076e-05, |
| "loss": 0.0576, |
| "step": 94000 |
| }, |
| { |
| "epoch": 235.0, |
| "eval_f1": 0.8940092165898618, |
| "eval_iou": 0.2772848308086395, |
| "eval_loss": 0.41021957993507385, |
| "eval_runtime": 19.2858, |
| "eval_samples_per_second": 4.563, |
| "eval_steps_per_second": 2.281, |
| "step": 94000 |
| }, |
| { |
| "epoch": 235.5, |
| "learning_rate": 1.8116760000000003e-05, |
| "loss": 0.0576, |
| "step": 94200 |
| }, |
| { |
| "epoch": 236.0, |
| "learning_rate": 1.811276e-05, |
| "loss": 0.0483, |
| "step": 94400 |
| }, |
| { |
| "epoch": 236.5, |
| "learning_rate": 1.8108760000000002e-05, |
| "loss": 0.0565, |
| "step": 94600 |
| }, |
| { |
| "epoch": 237.0, |
| "learning_rate": 1.810476e-05, |
| "loss": 0.071, |
| "step": 94800 |
| }, |
| { |
| "epoch": 237.5, |
| "learning_rate": 1.8100760000000002e-05, |
| "loss": 0.0576, |
| "step": 95000 |
| }, |
| { |
| "epoch": 238.0, |
| "learning_rate": 1.8096760000000003e-05, |
| "loss": 0.0585, |
| "step": 95200 |
| }, |
| { |
| "epoch": 238.5, |
| "learning_rate": 1.809276e-05, |
| "loss": 0.0486, |
| "step": 95400 |
| }, |
| { |
| "epoch": 239.0, |
| "learning_rate": 1.808876e-05, |
| "loss": 0.0475, |
| "step": 95600 |
| }, |
| { |
| "epoch": 239.5, |
| "learning_rate": 1.8084780000000002e-05, |
| "loss": 0.0708, |
| "step": 95800 |
| }, |
| { |
| "epoch": 240.0, |
| "learning_rate": 1.8080780000000003e-05, |
| "loss": 0.0593, |
| "step": 96000 |
| }, |
| { |
| "epoch": 240.0, |
| "eval_f1": 0.8906976744186047, |
| "eval_iou": 0.25792551040649414, |
| "eval_loss": 0.39649030566215515, |
| "eval_runtime": 19.5328, |
| "eval_samples_per_second": 4.505, |
| "eval_steps_per_second": 2.253, |
| "step": 96000 |
| }, |
| { |
| "epoch": 240.5, |
| "learning_rate": 1.807678e-05, |
| "loss": 0.0519, |
| "step": 96200 |
| }, |
| { |
| "epoch": 241.0, |
| "learning_rate": 1.807278e-05, |
| "loss": 0.0663, |
| "step": 96400 |
| }, |
| { |
| "epoch": 241.5, |
| "learning_rate": 1.806878e-05, |
| "loss": 0.0559, |
| "step": 96600 |
| }, |
| { |
| "epoch": 242.0, |
| "learning_rate": 1.8064780000000002e-05, |
| "loss": 0.0545, |
| "step": 96800 |
| }, |
| { |
| "epoch": 242.5, |
| "learning_rate": 1.8060780000000004e-05, |
| "loss": 0.0467, |
| "step": 97000 |
| }, |
| { |
| "epoch": 243.0, |
| "learning_rate": 1.805678e-05, |
| "loss": 0.0635, |
| "step": 97200 |
| }, |
| { |
| "epoch": 243.5, |
| "learning_rate": 1.805278e-05, |
| "loss": 0.0645, |
| "step": 97400 |
| }, |
| { |
| "epoch": 244.0, |
| "learning_rate": 1.804878e-05, |
| "loss": 0.0403, |
| "step": 97600 |
| }, |
| { |
| "epoch": 244.5, |
| "learning_rate": 1.8044780000000002e-05, |
| "loss": 0.0548, |
| "step": 97800 |
| }, |
| { |
| "epoch": 245.0, |
| "learning_rate": 1.8040780000000004e-05, |
| "loss": 0.0432, |
| "step": 98000 |
| }, |
| { |
| "epoch": 245.0, |
| "eval_f1": 0.9024943310657596, |
| "eval_iou": 0.2930222749710083, |
| "eval_loss": 0.3335016071796417, |
| "eval_runtime": 18.6665, |
| "eval_samples_per_second": 4.714, |
| "eval_steps_per_second": 2.357, |
| "step": 98000 |
| }, |
| { |
| "epoch": 245.5, |
| "learning_rate": 1.8036780000000002e-05, |
| "loss": 0.0589, |
| "step": 98200 |
| }, |
| { |
| "epoch": 246.0, |
| "learning_rate": 1.803278e-05, |
| "loss": 0.0616, |
| "step": 98400 |
| }, |
| { |
| "epoch": 246.5, |
| "learning_rate": 1.802878e-05, |
| "loss": 0.0463, |
| "step": 98600 |
| }, |
| { |
| "epoch": 247.0, |
| "learning_rate": 1.8024780000000003e-05, |
| "loss": 0.0632, |
| "step": 98800 |
| }, |
| { |
| "epoch": 247.5, |
| "learning_rate": 1.802078e-05, |
| "loss": 0.0524, |
| "step": 99000 |
| }, |
| { |
| "epoch": 248.0, |
| "learning_rate": 1.8016780000000002e-05, |
| "loss": 0.0602, |
| "step": 99200 |
| }, |
| { |
| "epoch": 248.5, |
| "learning_rate": 1.801278e-05, |
| "loss": 0.0618, |
| "step": 99400 |
| }, |
| { |
| "epoch": 249.0, |
| "learning_rate": 1.800878e-05, |
| "loss": 0.048, |
| "step": 99600 |
| }, |
| { |
| "epoch": 249.5, |
| "learning_rate": 1.8004780000000003e-05, |
| "loss": 0.0612, |
| "step": 99800 |
| }, |
| { |
| "epoch": 250.0, |
| "learning_rate": 1.800078e-05, |
| "loss": 0.0478, |
| "step": 100000 |
| }, |
| { |
| "epoch": 250.0, |
| "eval_f1": 0.8767772511848342, |
| "eval_iou": 0.26293623447418213, |
| "eval_loss": 0.37569504976272583, |
| "eval_runtime": 18.1469, |
| "eval_samples_per_second": 4.849, |
| "eval_steps_per_second": 2.425, |
| "step": 100000 |
| } |
| ], |
| "logging_steps": 200, |
| "max_steps": 1000000, |
| "num_train_epochs": 2500, |
| "save_steps": 2000, |
| "total_flos": 0.0, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|