raidavid commited on
Commit
e203f1f
·
1 Parent(s): 6db561a

Upload model on 2024-01-26 14:55:22.653813 epochs:500 model_nameyolov8n-seg imgsz:640

Browse files
BoxF1_curve.png CHANGED
BoxPR_curve.png CHANGED
BoxP_curve.png CHANGED
BoxR_curve.png CHANGED
MaskF1_curve.png CHANGED
MaskPR_curve.png CHANGED
MaskP_curve.png CHANGED
MaskR_curve.png CHANGED
args.yaml CHANGED
@@ -2,7 +2,7 @@ task: segment
2
  mode: train
3
  model: yolov8n-seg.pt
4
  data: data.yaml
5
- epochs: 1
6
  time: null
7
  patience: 50
8
  batch: 16
@@ -13,7 +13,7 @@ cache: false
13
  device: mps
14
  workers: 8
15
  project: null
16
- name: train21
17
  exist_ok: false
18
  pretrained: true
19
  optimizer: auto
@@ -98,4 +98,4 @@ mixup: 0.0
98
  copy_paste: 0.0
99
  cfg: null
100
  tracker: botsort.yaml
101
- save_dir: /Users/davidyang/code/tools/ultralytics/runs/segment/train21
 
2
  mode: train
3
  model: yolov8n-seg.pt
4
  data: data.yaml
5
+ epochs: 500
6
  time: null
7
  patience: 50
8
  batch: 16
 
13
  device: mps
14
  workers: 8
15
  project: null
16
+ name: train22
17
  exist_ok: false
18
  pretrained: true
19
  optimizer: auto
 
98
  copy_paste: 0.0
99
  cfg: null
100
  tracker: botsort.yaml
101
+ save_dir: /Users/davidyang/code/tools/ultralytics/runs/segment/train22
labels_correlogram.jpg CHANGED
results.csv CHANGED
@@ -1,2 +1,54 @@
1
  epoch, train/box_loss, train/seg_loss, train/cls_loss, train/dfl_loss, metrics/precision(B), metrics/recall(B), metrics/mAP50(B), metrics/mAP50-95(B), metrics/precision(M), metrics/recall(M), metrics/mAP50(M), metrics/mAP50-95(M), val/box_loss, val/seg_loss, val/cls_loss, val/dfl_loss, lr/pg0, lr/pg1, lr/pg2
2
- 1, 1.1662, 4.0717, 2.7859, 1.4613, 0.00325, 0.975, 0.61537, 0.39748, 0.00325, 0.975, 0.52803, 0.30489, 1.3798, 0.45842, 7.4438, 1.147, 0.00016, 0.00016, 0.00016
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  epoch, train/box_loss, train/seg_loss, train/cls_loss, train/dfl_loss, metrics/precision(B), metrics/recall(B), metrics/mAP50(B), metrics/mAP50-95(B), metrics/precision(M), metrics/recall(M), metrics/mAP50(M), metrics/mAP50-95(M), val/box_loss, val/seg_loss, val/cls_loss, val/dfl_loss, lr/pg0, lr/pg1, lr/pg2
2
+ 1, 1.3049, 4.3861, 2.8766, 1.5966, 0.00325, 0.975, 0.48338, 0.3264, 0.003, 0.9, 0.45012, 0.27107, 1.4977, 0.51025, 7.4335, 1.3306, 0.00016, 0.00016, 0.00016
3
+ 2, 1.0308, 3.306, 2.1512, 1.3505, 0.00333, 1, 0.56111, 0.33712, 0.00317, 0.95, 0.45509, 0.22203, 1.2447, 0.49035, 7.2099, 1.2104, 0.00033933, 0.00033933, 0.00033933
4
+ 3, 0.96073, 2.8103, 1.9436, 1.2902, 0.00325, 0.975, 0.66948, 0.4313, 0.00317, 0.95, 0.60425, 0.31202, 1.1185, 0.43721, 6.8853, 1.089, 0.00051794, 0.00051794, 0.00051794
5
+ 4, 1.0611, 2.7811, 1.8782, 1.3451, 0.83448, 0.25224, 0.71046, 0.40481, 0.7482, 0.22302, 0.52718, 0.27483, 1.418, 0.51561, 7.1094, 1.3092, 0.00069584, 0.00069584, 0.00069584
6
+ 5, 1.1542, 2.7605, 1.8499, 1.4255, 0.71481, 0.5, 0.69157, 0.35875, 0.54802, 0.375, 0.4096, 0.15805, 1.9653, 0.65516, 7.1002, 2.077, 0.00087303, 0.00087303, 0.00087303
7
+ 6, 1.1189, 2.5478, 1.7044, 1.3697, 0.58144, 0.425, 0.46999, 0.23572, 0.39979, 0.35, 0.43544, 0.19076, 1.7392, 0.42368, 7.4575, 1.5451, 0.0010495, 0.0010495, 0.0010495
8
+ 7, 1.1498, 2.6421, 1.7609, 1.4233, 0.16285, 0.2, 0.09336, 0.03361, 0.11757, 0.15, 0.04548, 0.0164, 2.803, 0.80774, 10.82, 2.7979, 0.0012253, 0.0012253, 0.0012253
9
+ 8, 1.255, 2.8755, 1.8716, 1.5203, 0.48724, 0.175, 0.12752, 0.05498, 0.36311, 0.15, 0.10502, 0.04173, 1.6564, 0.56905, 8.7045, 1.8441, 0.0014003, 0.0014003, 0.0014003
10
+ 9, 1.1665, 2.8394, 1.7073, 1.4156, 0.15813, 0.125, 0.0579, 0.03659, 0.03027, 0.025, 0.00104, 0.0001, 1.5016, 0.80119, 13.386, 1.8484, 0.0015747, 0.0015747, 0.0015747
11
+ 10, 1.2172, 2.8996, 1.6547, 1.4816, 0.19635, 0.125, 0.04898, 0.0236, 0.19635, 0.125, 0.04869, 0.01422, 1.725, 0.5209, 8.2782, 1.8099, 0.0017483, 0.0017483, 0.0017483
12
+ 11, 1.2161, 2.9078, 1.6949, 1.4928, 0.08306, 0.1, 0.01583, 0.01056, 0.06229, 0.075, 0.00908, 0.00269, 1.8445, 0.59372, 8.6799, 1.7876, 0.0019212, 0.0019212, 0.0019212
13
+ 12, 1.3092, 2.9519, 1.7542, 1.4779, 0.11464, 0.125, 0.02889, 0.00739, 0.12593, 0.075, 0.01927, 0.00769, 2.7395, 0.81151, 13.501, 2.2045, 0.0019564, 0.0019564, 0.0019564
14
+ 13, 1.4181, 3.1693, 1.797, 1.6035, 0.31053, 0.075, 0.04787, 0.0152, 0.21002, 0.05, 0.03666, 0.00916, 1.97, 0.70787, 13.537, 2.0801, 0.0019564, 0.0019564, 0.0019564
15
+ 14, 1.3136, 2.8019, 1.7488, 1.5275, 0.51902, 0.125, 0.10483, 0.06728, 0.51902, 0.125, 0.10399, 0.0306, 1.4391, 0.60718, 9.0405, 1.5956, 0.0019525, 0.0019525, 0.0019525
16
+ 15, 1.2046, 2.7595, 1.5403, 1.4688, 0.33204, 0.125, 0.09451, 0.04804, 0.33204, 0.125, 0.06844, 0.03383, 1.7366, 0.48681, 11.32, 1.6782, 0.0019485, 0.0019485, 0.0019485
17
+ 16, 1.2919, 2.8932, 1.5918, 1.5329, 0.15125, 0.075, 0.02199, 0.01394, 0.15125, 0.075, 0.02437, 0.01412, 1.7814, 0.64331, 16.225, 1.9027, 0.0019446, 0.0019446, 0.0019446
18
+ 17, 1.1947, 2.8454, 1.538, 1.4952, 0.22257, 0.1, 0.06922, 0.0378, 0.36142, 0.1, 0.07252, 0.03125, 1.5641, 0.50777, 12.884, 1.5221, 0.0019406, 0.0019406, 0.0019406
19
+ 18, 1.2474, 2.9649, 1.6675, 1.4822, 0.19753, 0.075, 0.0411, 0.01953, 0.19753, 0.075, 0.0411, 0.01479, 1.7122, 0.62054, 8.8629, 1.4649, 0.0019366, 0.0019366, 0.0019366
20
+ 19, 1.1501, 2.7646, 1.4091, 1.3956, 0.15484, 0.1, 0.04662, 0.01343, 0.11613, 0.075, 0.03733, 0.01116, 1.5921, 0.77809, 12.474, 1.4072, 0.0019327, 0.0019327, 0.0019327
21
+ 20, 1.1523, 2.9451, 1.4479, 1.4261, 0.13776, 0.025, 0.00885, 0.00463, 0.13776, 0.025, 0.00827, 0.00545, 1.6721, 0.67827, 11.009, 1.6782, 0.0019287, 0.0019287, 0.0019287
22
+ 21, 1.1983, 2.9708, 1.4819, 1.43, 0.05423, 0.05, 0.01871, 0.00497, 0.05883, 0.05, 0.01829, 0.00497, 1.6161, 0.56584, 8.2173, 1.3964, 0.0019248, 0.0019248, 0.0019248
23
+ 22, 1.3018, 2.9762, 1.512, 1.5226, 0.27994, 0.1, 0.06532, 0.04206, 0.27994, 0.1, 0.06532, 0.04472, 1.7146, 0.51666, 5.1646, 1.4218, 0.0019208, 0.0019208, 0.0019208
24
+ 23, 1.266, 2.8495, 1.4139, 1.4483, 0.95285, 0.075, 0.07685, 0.05032, 0.95285, 0.075, 0.07685, 0.04891, 1.6163, 0.42382, 4.0061, 1.3454, 0.0019168, 0.0019168, 0.0019168
25
+ 24, 1.2469, 2.8688, 1.409, 1.4964, 0.97417, 0.075, 0.07674, 0.0627, 0.97417, 0.075, 0.07674, 0.05023, 1.4727, 0.42196, 3.7573, 1.2754, 0.0019129, 0.0019129, 0.0019129
26
+ 25, 1.1458, 2.9977, 1.3678, 1.4098, 0.57648, 0.075, 0.06454, 0.04763, 0.57648, 0.075, 0.06454, 0.04369, 1.6688, 0.67848, 5.0254, 1.4159, 0.0019089, 0.0019089, 0.0019089
27
+ 26, 1.1834, 2.7924, 1.2971, 1.3988, 0.4384, 0.075, 0.06175, 0.04574, 0.4384, 0.075, 0.06175, 0.04139, 1.4519, 0.71703, 3.9818, 1.31, 0.001905, 0.001905, 0.001905
28
+ 27, 1.0771, 2.7104, 1.2472, 1.3594, 0.70324, 0.075, 0.06938, 0.04452, 0.70324, 0.075, 0.06938, 0.0492, 1.4846, 0.40857, 4.5511, 1.3521, 0.001901, 0.001901, 0.001901
29
+ 28, 1.2085, 2.8933, 1.3974, 1.4679, 0.90075, 0.075, 0.07685, 0.04725, 0.90075, 0.075, 0.07685, 0.05023, 1.9156, 0.44637, 5.633, 1.6539, 0.001897, 0.001897, 0.001897
30
+ 29, 1.1798, 2.5898, 1.3279, 1.4428, 0.78427, 0.05, 0.04799, 0.04009, 0.14874, 0.075, 0.05203, 0.02826, 2.0564, 0.59022, 5.1429, 1.9787, 0.0018931, 0.0018931, 0.0018931
31
+ 30, 1.1327, 2.9491, 1.2692, 1.3953, 0.78784, 0.075, 0.0769, 0.03945, 0.78784, 0.075, 0.0769, 0.01863, 1.756, 0.66536, 3.7712, 1.5563, 0.0018891, 0.0018891, 0.0018891
32
+ 31, 1.2039, 2.8944, 1.359, 1.464, 0.89693, 0.075, 0.07705, 0.04415, 0.89693, 0.075, 0.07705, 0.04493, 1.5349, 0.44924, 3.545, 1.4001, 0.0018852, 0.0018852, 0.0018852
33
+ 32, 1.1774, 2.8228, 1.326, 1.4092, 0.58952, 0.075, 0.07011, 0.04124, 0.58952, 0.075, 0.07011, 0.03543, 1.9599, 0.51646, 4.8026, 1.6926, 0.0018812, 0.0018812, 0.0018812
34
+ 33, 1.2779, 2.7124, 1.3498, 1.4709, 0.72109, 0.075, 0.06433, 0.03959, 0.72109, 0.075, 0.06433, 0.03218, 1.5953, 0.39348, 3.4074, 1.3985, 0.0018772, 0.0018772, 0.0018772
35
+ 34, 1.107, 2.8973, 1.2006, 1.3641, 0.72429, 0.075, 0.07669, 0.04635, 0.72429, 0.075, 0.07669, 0.038, 1.4612, 0.44161, 3.4556, 1.4207, 0.0018733, 0.0018733, 0.0018733
36
+ 35, 1.0998, 2.7443, 1.1853, 1.3776, 0.32482, 0.075, 0.03875, 0.02682, 0.32482, 0.075, 0.03554, 0.02126, 1.4156, 0.38351, 3.2765, 1.3928, 0.0018693, 0.0018693, 0.0018693
37
+ 36, 1.1672, 2.9225, 1.2452, 1.4592, 0.71353, 0.075, 0.07056, 0.05311, 0.71353, 0.075, 0.07056, 0.05791, 1.3422, 0.34072, 2.832, 1.4048, 0.0018654, 0.0018654, 0.0018654
38
+ 37, 1.1275, 2.7985, 1.2976, 1.4003, 0.62857, 0.075, 0.06529, 0.05223, 0.62857, 0.075, 0.06529, 0.05223, 1.1609, 0.35011, 2.5066, 1.1034, 0.0018614, 0.0018614, 0.0018614
39
+ 38, 1.0685, 2.5661, 1.2099, 1.331, 0.67175, 0.075, 0.0718, 0.05646, 0.67175, 0.075, 0.0718, 0.05443, 1.1974, 0.35292, 2.8653, 1.1056, 0.0018574, 0.0018574, 0.0018574
40
+ 39, 1.0386, 2.7696, 1.1386, 1.3207, 0.65706, 0.1, 0.08044, 0.06281, 0.85886, 0.1, 0.09944, 0.06561, 1.4896, 0.32558, 4.3823, 1.2739, 0.0018535, 0.0018535, 0.0018535
41
+ 40, 1.1131, 2.7157, 1.186, 1.362, 0.91141, 0.1, 0.09737, 0.05504, 0.91141, 0.1, 0.09737, 0.06349, 1.533, 0.34255, 4.1234, 1.3965, 0.0018495, 0.0018495, 0.0018495
42
+ 41, 1.0318, 2.6202, 1.105, 1.3452, 0.94642, 0.05, 0.04774, 0.02215, 0.94642, 0.05, 0.04774, 0.0136, 1.65, 0.4342, 4.2352, 1.548, 0.0018456, 0.0018456, 0.0018456
43
+ 42, 1.1188, 2.8251, 1.1798, 1.4131, 0.9746, 0.05, 0.04951, 0.03738, 0.9746, 0.05, 0.04951, 0.04044, 1.5216, 0.34241, 3.6841, 1.3558, 0.0018416, 0.0018416, 0.0018416
44
+ 43, 1.05, 2.5349, 1.1626, 1.3428, 0.92011, 0.05, 0.05025, 0.03517, 0.92011, 0.05, 0.05025, 0.03366, 1.5579, 0.36236, 3.6104, 1.4536, 0.0018376, 0.0018376, 0.0018376
45
+ 44, 1.0589, 2.6585, 1.1245, 1.3274, 0.96379, 0.05, 0.04875, 0.0302, 0.96379, 0.05, 0.04908, 0.02737, 1.7412, 0.42511, 4.759, 1.4644, 0.0018337, 0.0018337, 0.0018337
46
+ 45, 1.0552, 2.6133, 1.0887, 1.3298, 0.90362, 0.05, 0.04693, 0.04014, 0.90362, 0.05, 0.04693, 0.02423, 1.7257, 0.45363, 4.8935, 1.4663, 0.0018297, 0.0018297, 0.0018297
47
+ 46, 1.0374, 2.5344, 1.0032, 1.3207, 0.63032, 0.05, 0.04675, 0.03053, 0.10495, 0.03266, 0.02983, 0.01863, 1.5134, 0.51293, 4.2263, 1.3124, 0.0018258, 0.0018258, 0.0018258
48
+ 47, 1.0477, 2.4975, 1.1085, 1.2972, 0.92235, 0.05, 0.04829, 0.03796, 0.92235, 0.05, 0.04829, 0.0353, 1.166, 0.46474, 3.4198, 1.0978, 0.0018218, 0.0018218, 0.0018218
49
+ 48, 1.0463, 2.5151, 1.0309, 1.3122, 0.94615, 0.075, 0.07992, 0.06536, 0.94615, 0.075, 0.07992, 0.06078, 1.3313, 0.62526, 2.626, 1.2188, 0.0018178, 0.0018178, 0.0018178
50
+ 49, 1.0766, 2.4695, 1.1812, 1.3441, 0.95449, 0.05, 0.049, 0.03721, 0.95449, 0.05, 0.049, 0.04065, 1.2112, 0.56308, 2.5554, 1.1316, 0.0018139, 0.0018139, 0.0018139
51
+ 50, 0.99325, 2.507, 1.0551, 1.3077, 0.95424, 0.05, 0.05025, 0.03845, 0.95424, 0.05, 0.05025, 0.0407, 1.1327, 0.52174, 2.5609, 1.0804, 0.0018099, 0.0018099, 0.0018099
52
+ 51, 0.9591, 2.4474, 1.0563, 1.274, 0.61961, 0.05, 0.04461, 0.02576, 0.61961, 0.05, 0.04461, 0.01786, 1.1636, 0.47, 2.6062, 1.094, 0.001806, 0.001806, 0.001806
53
+ 52, 1.0944, 2.5755, 1.0922, 1.3553, 0.46381, 0.05, 0.04134, 0.03274, 0.39824, 0.075, 0.05391, 0.04413, 1.7657, 0.41693, 4.7234, 1.7854, 0.001802, 0.001802, 0.001802
54
+ 53, 1.0697, 2.8319, 1.1175, 1.2994, 0.85086, 0.025, 0.02884, 0.01937, 0.85086, 0.025, 0.02884, 0.02446, 1.9206, 0.53429, 5.4211, 2.0018, 0.001798, 0.001798, 0.001798
results.png CHANGED
train_batch0.jpg CHANGED
train_batch1.jpg CHANGED
val_batch1_pred.jpg CHANGED
weights/best.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49a3490359b0a2d293c2243443130a05890b9f1e293c545d0c54b07aafb74ac9
3
- size 6775651
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:025960c1e10646ca1667d11e65c8c25cad9013e488c301ae7ee8cb7cef7f59d4
3
+ size 6776035
weights/last.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95c01297647c1a47cdaf73084e2fc34d5e617e48313fe5a385f48f7143ce55e9
3
- size 6775651
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa79166a2aab33c15004a5c9f7dc543c7bd87be4dc36de6a9fceb63641ff840
3
+ size 6784675