Drazcat-AI commited on
Commit
bf6e268
·
verified ·
1 Parent(s): da823f8

Delete blender_data-29

Browse files
blender_data-29/runs/detect/train/F1_curve.png DELETED
Binary file (659 kB)
 
blender_data-29/runs/detect/train/PR_curve.png DELETED
Binary file (544 kB)
 
blender_data-29/runs/detect/train/P_curve.png DELETED
Binary file (667 kB)
 
blender_data-29/runs/detect/train/R_curve.png DELETED
Binary file (555 kB)
 
blender_data-29/runs/detect/train/args.yaml DELETED
@@ -1,98 +0,0 @@
1
- task: detect
2
- mode: train
3
- model: yolov8m.pt
4
- data: /home/ec2-user/SageMaker/datasets/blender_data-29/data.yaml
5
- epochs: 300
6
- patience: 50
7
- batch: 4
8
- imgsz: 1280
9
- save: true
10
- save_period: -1
11
- cache: false
12
- device: null
13
- workers: 8
14
- project: null
15
- name: null
16
- exist_ok: false
17
- pretrained: true
18
- optimizer: auto
19
- verbose: true
20
- seed: 0
21
- deterministic: true
22
- single_cls: false
23
- rect: false
24
- cos_lr: false
25
- close_mosaic: 10
26
- resume: false
27
- amp: true
28
- fraction: 1.0
29
- profile: false
30
- freeze: null
31
- overlap_mask: true
32
- mask_ratio: 4
33
- dropout: 0.0
34
- val: true
35
- split: val
36
- save_json: false
37
- save_hybrid: false
38
- conf: null
39
- iou: 0.7
40
- max_det: 300
41
- half: false
42
- dnn: false
43
- plots: true
44
- source: null
45
- show: false
46
- save_txt: false
47
- save_conf: false
48
- save_crop: false
49
- show_labels: true
50
- show_conf: true
51
- vid_stride: 1
52
- stream_buffer: false
53
- line_width: null
54
- visualize: false
55
- augment: false
56
- agnostic_nms: false
57
- classes: null
58
- retina_masks: false
59
- boxes: true
60
- format: torchscript
61
- keras: false
62
- optimize: false
63
- int8: false
64
- dynamic: false
65
- simplify: false
66
- opset: null
67
- workspace: 4
68
- nms: false
69
- lr0: 0.01
70
- lrf: 0.01
71
- momentum: 0.937
72
- weight_decay: 0.0005
73
- warmup_epochs: 3.0
74
- warmup_momentum: 0.8
75
- warmup_bias_lr: 0.1
76
- box: 7.5
77
- cls: 0.5
78
- dfl: 1.5
79
- pose: 12.0
80
- kobj: 1.0
81
- label_smoothing: 0.0
82
- nbs: 64
83
- hsv_h: 0.015
84
- hsv_s: 0.7
85
- hsv_v: 0.4
86
- degrees: 0.0
87
- translate: 0.1
88
- scale: 0.5
89
- shear: 0.0
90
- perspective: 0.0
91
- flipud: 0.0
92
- fliplr: 0.5
93
- mosaic: 1.0
94
- mixup: 0.0
95
- copy_paste: 0.0
96
- cfg: null
97
- tracker: botsort.yaml
98
- save_dir: runs/detect/train
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
blender_data-29/runs/detect/train/confusion_matrix.png DELETED
Binary file (799 kB)
 
blender_data-29/runs/detect/train/confusion_matrix_normalized.png DELETED
Binary file (825 kB)
 
blender_data-29/runs/detect/train/labels.jpg DELETED
Binary file (233 kB)
 
blender_data-29/runs/detect/train/labels_correlogram.jpg DELETED
Binary file (255 kB)
 
blender_data-29/runs/detect/train/model_artifacts.json DELETED
@@ -1 +0,0 @@
1
- {"names": ["AVENA TRADICIONAL QUAKER 700 GRS -ID-", "Carozzi Fideos Cabellos de angel Bolsa 400 g -ID- 23", "Copihue Pack Cajas Fosforos 10 Un 38 Un aprox cu -ID- 811", "Harina Sin Polvos Selecta 1 Kg -ID-", "Ketchup Heinz 900ml -ID-", "Lavaloza Quix Bio-Activos Limon 750 Ml -ID-", "Leche Entera Soprole 1 Lt -ID-", "Lucchetti Fideos Espirales Trisabor Bolsa 400 g -ID- 536", "Mildred Te Ceylan Caja 200 g -ID- 704", "SAL DE MESA LOBOS 1 KG -ID-", "San Jose Jurel Natural Abre Facil 425 g -ID- 225", "Suavizante Soft Liquido Classic 1 L -ID- 1672", "Sucedaneo De Limon Traverso 500 Cc -ID-", "Supremo Te Ceylan Premium Caja 100 Un -ID- 146", "Triton Galletas Triton Naranja 126 g -ID- 312", "Triton Galletas Triton Vainilla 126 g -ID- 94", "Tucapel Arroz Grado 2 Seleccion Bolsa 1 kg -ID- 6", "Watts Nectar Sabor Tutti-Frutilla 1 1_5 L -ID- 108", "Watts Nectar de Damasco 1_5 L -ID- 360", "Watts Nectar de Pina 1-5 L -ID- 366", "Yoghurt batido Colun Vainilla 125gr -ID-"], "yaml": {"nc": 21, "depth_multiple": 0.67, "width_multiple": 0.75, "backbone": [[-1, 1, "Conv", [64, 3, 2]], [-1, 1, "Conv", [128, 3, 2]], [-1, 3, "C2f", [128, true]], [-1, 1, "Conv", [256, 3, 2]], [-1, 6, "C2f", [256, true]], [-1, 1, "Conv", [512, 3, 2]], [-1, 6, "C2f", [512, true]], [-1, 1, "Conv", [768, 3, 2]], [-1, 3, "C2f", [768, true]], [-1, 1, "SPPF", [768, 5]]], "head": [[-1, 1, "nn.Upsample", ["None", 2, "nearest"]], [[-1, 6], 1, "Concat", [1]], [-1, 3, "C2f", [512]], [-1, 1, "nn.Upsample", ["None", 2, "nearest"]], [[-1, 4], 1, "Concat", [1]], [-1, 3, "C2f", [256]], [-1, 1, "Conv", [256, 3, 2]], [[-1, 12], 1, "Concat", [1]], [-1, 3, "C2f", [512]], [-1, 1, "Conv", [512, 3, 2]], [[-1, 9], 1, "Concat", [1]], [-1, 3, "C2f", [768]], [[15, 18, 21], 1, "Detect", ["nc"]]], "ch": 3}, "nc": 21, "args": {"model": "yolov8m.pt", "batch": 4, "imgsz": 1280}, "ultralytics_version": "8.0.196", "model_type": "yolov8"}
 
 
blender_data-29/runs/detect/train/results.csv DELETED
@@ -1,99 +0,0 @@
1
- epoch, train/box_loss, train/cls_loss, train/dfl_loss, metrics/precision(B), metrics/recall(B), metrics/mAP50(B), metrics/mAP50-95(B), val/box_loss, val/cls_loss, val/dfl_loss, lr/pg0, lr/pg1, lr/pg2
2
- 1, 1.3662, 2.6141, 1.3593, 0.24504, 0.40346, 0.26465, 0.16622, 1.6298, 17.779, 1.566, 0.00013162, 0.00013162, 0.00013162
3
- 2, 0.92348, 0.98322, 1.1044, 0.28894, 0.45049, 0.34388, 0.21612, 1.7758, 8.7101, 1.7578, 0.00026408, 0.00026408, 0.00026408
4
- 3, 0.94485, 0.84505, 1.1217, 0.27385, 0.4027, 0.29365, 0.17534, 1.8949, 8.4714, 1.8528, 0.00039566, 0.00039566, 0.00039566
5
- 4, 0.84056, 0.73123, 1.0826, 0.3579, 0.39831, 0.34404, 0.20962, 1.9821, 6.4445, 1.9299, 0.00039604, 0.00039604, 0.00039604
6
- 5, 0.76013, 0.62345, 1.0188, 0.37126, 0.44904, 0.35178, 0.20754, 1.9223, 5.8656, 1.9059, 0.00039604, 0.00039604, 0.00039604
7
- 6, 0.74663, 0.59659, 1.0232, 0.34149, 0.43771, 0.34328, 0.20799, 1.8289, 6.3923, 1.8809, 0.00039472, 0.00039472, 0.00039472
8
- 7, 0.70927, 0.57006, 0.98972, 0.3177, 0.38445, 0.32346, 0.19686, 1.8023, 5.3868, 1.8105, 0.0003934, 0.0003934, 0.0003934
9
- 8, 0.72338, 0.53886, 1.0023, 0.32778, 0.50846, 0.3817, 0.24124, 1.7943, 5.4297, 1.8135, 0.00039208, 0.00039208, 0.00039208
10
- 9, 0.65347, 0.49315, 0.96753, 0.32523, 0.52836, 0.39371, 0.24314, 1.9135, 3.9409, 1.9262, 0.00039076, 0.00039076, 0.00039076
11
- 10, 0.66025, 0.46786, 0.96526, 0.38187, 0.47928, 0.38504, 0.233, 1.9581, 4.0373, 1.9858, 0.00038944, 0.00038944, 0.00038944
12
- 11, 0.64829, 0.47869, 0.94973, 0.31068, 0.46207, 0.34326, 0.21423, 1.8846, 4.9197, 1.9457, 0.00038812, 0.00038812, 0.00038812
13
- 12, 0.6052, 0.43254, 0.93561, 0.32926, 0.45536, 0.31731, 0.18372, 1.884, 4.4412, 1.9437, 0.0003868, 0.0003868, 0.0003868
14
- 13, 0.54713, 0.40774, 0.91552, 0.30544, 0.46821, 0.29995, 0.17471, 1.9346, 4.5383, 2.0012, 0.00038548, 0.00038548, 0.00038548
15
- 14, 0.58482, 0.43111, 0.93208, 0.28471, 0.49081, 0.3024, 0.18747, 1.8763, 4.4775, 1.9134, 0.00038416, 0.00038416, 0.00038416
16
- 15, 0.57559, 0.4125, 0.91982, 0.28297, 0.51594, 0.29927, 0.17963, 1.8893, 4.2666, 1.9591, 0.00038284, 0.00038284, 0.00038284
17
- 16, 0.56763, 0.40316, 0.91579, 0.28915, 0.50654, 0.32345, 0.1987, 1.9303, 4.5303, 1.9953, 0.00038152, 0.00038152, 0.00038152
18
- 17, 0.54647, 0.39734, 0.91722, 0.36056, 0.44534, 0.33071, 0.20885, 1.9648, 4.3083, 2.0128, 0.0003802, 0.0003802, 0.0003802
19
- 18, 0.52145, 0.37863, 0.89727, 0.31232, 0.50248, 0.33579, 0.20119, 1.9336, 4.635, 2.0764, 0.00037888, 0.00037888, 0.00037888
20
- 19, 0.51799, 0.37126, 0.90145, 0.32919, 0.45944, 0.3243, 0.19201, 1.956, 4.4467, 2.0449, 0.00037756, 0.00037756, 0.00037756
21
- 20, 0.48508, 0.35543, 0.8795, 0.3383, 0.48181, 0.32879, 0.20066, 1.8275, 3.9914, 1.9264, 0.00037624, 0.00037624, 0.00037624
22
- 21, 0.5328, 0.38403, 0.90413, 0.36108, 0.48382, 0.3534, 0.21294, 1.9172, 4.546, 2.0178, 0.00037492, 0.00037492, 0.00037492
23
- 22, 0.53125, 0.3702, 0.90511, 0.34355, 0.52798, 0.36381, 0.21673, 1.8764, 4.6911, 2.0085, 0.0003736, 0.0003736, 0.0003736
24
- 23, 0.50431, 0.35284, 0.88728, 0.32465, 0.49129, 0.36631, 0.22825, 1.9089, 4.8394, 2.0342, 0.00037228, 0.00037228, 0.00037228
25
- 24, 0.49351, 0.35473, 0.88812, 0.37886, 0.49735, 0.38984, 0.24298, 1.8222, 4.4461, 1.9177, 0.00037096, 0.00037096, 0.00037096
26
- 25, 0.45371, 0.3377, 0.87385, 0.39003, 0.54552, 0.38051, 0.22603, 1.8444, 4.3442, 2.0218, 0.00036964, 0.00036964, 0.00036964
27
- 26, 0.47662, 0.34388, 0.8764, 0.33167, 0.55851, 0.3644, 0.22038, 1.8799, 4.5334, 2.076, 0.00036832, 0.00036832, 0.00036832
28
- 27, 0.47371, 0.33785, 0.88294, 0.34442, 0.45397, 0.32773, 0.20531, 1.9101, 4.6791, 2.0861, 0.000367, 0.000367, 0.000367
29
- 28, 0.44959, 0.3248, 0.86815, 0.3115, 0.48113, 0.33245, 0.1981, 1.9628, 4.6004, 2.1596, 0.00036568, 0.00036568, 0.00036568
30
- 29, 0.45127, 0.31689, 0.87503, 0.34131, 0.4913, 0.37097, 0.22838, 1.9186, 3.8991, 2.1136, 0.00036436, 0.00036436, 0.00036436
31
- 30, 0.47105, 0.33928, 0.87353, 0.32114, 0.47418, 0.38205, 0.23601, 1.9863, 3.6698, 2.1507, 0.00036304, 0.00036304, 0.00036304
32
- 31, 0.46181, 0.33744, 0.88182, 0.36958, 0.43965, 0.36424, 0.22714, 2.0002, 4.5664, 2.1478, 0.00036172, 0.00036172, 0.00036172
33
- 32, 0.46017, 0.33185, 0.874, 0.3987, 0.45192, 0.3953, 0.24739, 1.9508, 4.8316, 2.1381, 0.0003604, 0.0003604, 0.0003604
34
- 33, 0.43425, 0.31959, 0.86654, 0.39427, 0.49044, 0.42498, 0.253, 1.8966, 3.5673, 2.0902, 0.00035908, 0.00035908, 0.00035908
35
- 34, 0.4186, 0.30067, 0.86096, 0.40594, 0.5147, 0.41621, 0.24738, 1.8918, 3.6185, 2.1083, 0.00035776, 0.00035776, 0.00035776
36
- 35, 0.41925, 0.30728, 0.86804, 0.37224, 0.50546, 0.38742, 0.23443, 1.8823, 3.3333, 2.083, 0.00035644, 0.00035644, 0.00035644
37
- 36, 0.39674, 0.29053, 0.85553, 0.36254, 0.48488, 0.36537, 0.22136, 1.9267, 3.4886, 2.0962, 0.00035512, 0.00035512, 0.00035512
38
- 37, 0.41963, 0.30054, 0.86198, 0.44296, 0.45031, 0.3762, 0.23162, 1.9233, 3.6512, 2.1303, 0.0003538, 0.0003538, 0.0003538
39
- 38, 0.42702, 0.30246, 0.86194, 0.38436, 0.51956, 0.38409, 0.23647, 1.9077, 3.9738, 2.1335, 0.00035248, 0.00035248, 0.00035248
40
- 39, 0.42189, 0.2952, 0.85983, 0.39667, 0.47149, 0.37994, 0.22877, 1.9197, 3.6482, 2.1285, 0.00035116, 0.00035116, 0.00035116
41
- 40, 0.4248, 0.29848, 0.86235, 0.34794, 0.4932, 0.34861, 0.22121, 1.9197, 4.2368, 2.1022, 0.00034984, 0.00034984, 0.00034984
42
- 41, 0.41638, 0.28759, 0.85795, 0.4169, 0.46188, 0.35456, 0.22173, 1.8992, 3.9915, 2.0972, 0.00034852, 0.00034852, 0.00034852
43
- 42, 0.41193, 0.28837, 0.85065, 0.33361, 0.48242, 0.33825, 0.20568, 1.9133, 4.0126, 2.1362, 0.0003472, 0.0003472, 0.0003472
44
- 43, 0.38641, 0.27637, 0.84523, 0.37678, 0.50557, 0.37805, 0.23646, 1.8761, 3.9679, 2.0954, 0.00034588, 0.00034588, 0.00034588
45
- 44, 0.39164, 0.28081, 0.84869, 0.49194, 0.49253, 0.40148, 0.24954, 1.8148, 3.7416, 2.0058, 0.00034456, 0.00034456, 0.00034456
46
- 45, 0.39595, 0.2821, 0.84038, 0.35034, 0.47412, 0.35292, 0.21917, 1.9433, 3.7266, 2.1572, 0.00034324, 0.00034324, 0.00034324
47
- 46, 0.39405, 0.28258, 0.8392, 0.37195, 0.46419, 0.36944, 0.22526, 2.0058, 3.9923, 2.2657, 0.00034192, 0.00034192, 0.00034192
48
- 47, 0.40353, 0.28003, 0.84953, 0.37318, 0.51149, 0.39403, 0.24674, 1.889, 4.1941, 2.1715, 0.0003406, 0.0003406, 0.0003406
49
- 48, 0.39427, 0.27784, 0.8485, 0.39629, 0.5271, 0.42412, 0.26402, 1.8535, 3.6523, 2.1072, 0.00033928, 0.00033928, 0.00033928
50
- 49, 0.40969, 0.28869, 0.8521, 0.39596, 0.47925, 0.39999, 0.24937, 1.8924, 3.368, 2.1325, 0.00033796, 0.00033796, 0.00033796
51
- 50, 0.37979, 0.28187, 0.84737, 0.40014, 0.48808, 0.40063, 0.24692, 1.9014, 3.7108, 2.1939, 0.00033664, 0.00033664, 0.00033664
52
- 51, 0.38649, 0.27255, 0.84846, 0.38232, 0.48051, 0.38933, 0.238, 1.8988, 4.0838, 2.1585, 0.00033532, 0.00033532, 0.00033532
53
- 52, 0.38148, 0.26998, 0.84466, 0.42409, 0.48838, 0.4022, 0.24305, 1.8758, 3.8205, 2.099, 0.000334, 0.000334, 0.000334
54
- 53, 0.36665, 0.26156, 0.83376, 0.41538, 0.49452, 0.41001, 0.2479, 1.8914, 3.8177, 2.1549, 0.00033268, 0.00033268, 0.00033268
55
- 54, 0.38155, 0.26518, 0.83787, 0.36253, 0.47774, 0.35542, 0.21558, 1.9154, 4.1343, 2.1792, 0.00033136, 0.00033136, 0.00033136
56
- 55, 0.40991, 0.28471, 0.84389, 0.37731, 0.47534, 0.38127, 0.23861, 1.839, 3.7542, 2.127, 0.00033004, 0.00033004, 0.00033004
57
- 56, 0.38241, 0.27486, 0.84213, 0.4037, 0.43384, 0.40264, 0.24527, 1.9326, 3.6956, 2.1606, 0.00032872, 0.00032872, 0.00032872
58
- 57, 0.36413, 0.27082, 0.83676, 0.37779, 0.44038, 0.3697, 0.229, 1.9402, 3.6503, 2.1684, 0.0003274, 0.0003274, 0.0003274
59
- 58, 0.35755, 0.26156, 0.83199, 0.43413, 0.51501, 0.39136, 0.24427, 1.9307, 3.9593, 2.1549, 0.00032608, 0.00032608, 0.00032608
60
- 59, 0.35622, 0.25982, 0.83625, 0.40295, 0.51192, 0.39666, 0.23683, 1.9297, 3.9258, 2.2235, 0.00032476, 0.00032476, 0.00032476
61
- 60, 0.34257, 0.25046, 0.82731, 0.38054, 0.45849, 0.36584, 0.22889, 1.8877, 3.7227, 2.1168, 0.00032344, 0.00032344, 0.00032344
62
- 61, 0.34754, 0.25186, 0.83342, 0.33646, 0.48748, 0.36144, 0.21817, 1.8763, 3.6816, 2.1453, 0.00032212, 0.00032212, 0.00032212
63
- 62, 0.36455, 0.26512, 0.83428, 0.37533, 0.45948, 0.33497, 0.20403, 1.9328, 3.7235, 2.1974, 0.0003208, 0.0003208, 0.0003208
64
- 63, 0.35067, 0.25563, 0.83303, 0.38149, 0.40031, 0.32956, 0.20224, 1.976, 4.3167, 2.2395, 0.00031948, 0.00031948, 0.00031948
65
- 64, 0.34742, 0.25101, 0.82982, 0.32695, 0.48682, 0.35169, 0.21001, 1.9661, 4.4573, 2.2521, 0.00031816, 0.00031816, 0.00031816
66
- 65, 0.36759, 0.26408, 0.84315, 0.45254, 0.4313, 0.3498, 0.21013, 2.0295, 3.9646, 2.2983, 0.00031684, 0.00031684, 0.00031684
67
- 66, 0.34277, 0.24677, 0.83425, 0.37354, 0.44618, 0.35108, 0.20958, 2.0073, 3.5345, 2.2755, 0.00031552, 0.00031552, 0.00031552
68
- 67, 0.34672, 0.25191, 0.83416, 0.36368, 0.41007, 0.3548, 0.21493, 2.0116, 4.0481, 2.3402, 0.0003142, 0.0003142, 0.0003142
69
- 68, 0.33881, 0.24547, 0.83505, 0.38061, 0.45184, 0.35716, 0.21313, 2.0078, 3.6342, 2.3063, 0.00031288, 0.00031288, 0.00031288
70
- 69, 0.33177, 0.23826, 0.82183, 0.34939, 0.43448, 0.32796, 0.19444, 2.0648, 3.6867, 2.3585, 0.00031156, 0.00031156, 0.00031156
71
- 70, 0.35316, 0.2491, 0.82865, 0.34432, 0.38459, 0.32218, 0.18972, 2.0602, 3.6994, 2.3452, 0.00031024, 0.00031024, 0.00031024
72
- 71, 0.3475, 0.24473, 0.82905, 0.35061, 0.45614, 0.33889, 0.19789, 2.0707, 3.8044, 2.3847, 0.00030892, 0.00030892, 0.00030892
73
- 72, 0.36065, 0.24706, 0.8353, 0.35583, 0.46191, 0.33835, 0.20168, 1.9473, 3.9564, 2.2579, 0.0003076, 0.0003076, 0.0003076
74
- 73, 0.3457, 0.24493, 0.83443, 0.413, 0.46761, 0.36484, 0.21995, 1.9216, 3.7215, 2.2006, 0.00030628, 0.00030628, 0.00030628
75
- 74, 0.33913, 0.24071, 0.82409, 0.41735, 0.43612, 0.34349, 0.20403, 1.9998, 3.8952, 2.3068, 0.00030496, 0.00030496, 0.00030496
76
- 75, 0.33045, 0.23385, 0.82174, 0.34141, 0.47733, 0.33473, 0.20006, 1.9616, 4.076, 2.2509, 0.00030364, 0.00030364, 0.00030364
77
- 76, 0.32356, 0.23218, 0.82334, 0.38318, 0.44609, 0.36363, 0.21643, 1.9841, 3.7672, 2.2726, 0.00030232, 0.00030232, 0.00030232
78
- 77, 0.33336, 0.2363, 0.82342, 0.36669, 0.46747, 0.37235, 0.22405, 1.9658, 3.9411, 2.2339, 0.000301, 0.000301, 0.000301
79
- 78, 0.35609, 0.24529, 0.82998, 0.37956, 0.50407, 0.38378, 0.22466, 1.9109, 4.0312, 2.2064, 0.00029968, 0.00029968, 0.00029968
80
- 79, 0.34938, 0.24804, 0.8287, 0.46451, 0.46723, 0.38365, 0.22882, 1.9205, 3.5121, 2.1803, 0.00029836, 0.00029836, 0.00029836
81
- 80, 0.32867, 0.23208, 0.82483, 0.40285, 0.50413, 0.36622, 0.21936, 1.933, 3.3208, 2.2168, 0.00029704, 0.00029704, 0.00029704
82
- 81, 0.32611, 0.23216, 0.82549, 0.383, 0.46733, 0.38041, 0.22612, 1.9033, 3.4809, 2.1995, 0.00029572, 0.00029572, 0.00029572
83
- 82, 0.33795, 0.23674, 0.82298, 0.36263, 0.49409, 0.37846, 0.23017, 2.0153, 3.5829, 2.3238, 0.0002944, 0.0002944, 0.0002944
84
- 83, 0.32592, 0.22824, 0.82383, 0.45353, 0.49286, 0.40704, 0.24548, 1.8907, 3.632, 2.2556, 0.00029308, 0.00029308, 0.00029308
85
- 84, 0.31379, 0.22869, 0.82174, 0.37392, 0.5043, 0.38437, 0.24162, 1.8725, 3.8535, 2.2114, 0.00029176, 0.00029176, 0.00029176
86
- 85, 0.32175, 0.2287, 0.81924, 0.34644, 0.47316, 0.36489, 0.22679, 1.8866, 3.6094, 2.1893, 0.00029044, 0.00029044, 0.00029044
87
- 86, 0.3126, 0.2239, 0.82236, 0.35642, 0.47472, 0.36785, 0.22486, 1.9172, 3.4468, 2.2602, 0.00028912, 0.00028912, 0.00028912
88
- 87, 0.33796, 0.23948, 0.82638, 0.33096, 0.45933, 0.34692, 0.21011, 1.9556, 3.695, 2.2791, 0.0002878, 0.0002878, 0.0002878
89
- 88, 0.32398, 0.231, 0.8242, 0.3998, 0.43442, 0.3436, 0.20898, 1.9222, 3.8738, 2.2655, 0.00028648, 0.00028648, 0.00028648
90
- 89, 0.32953, 0.23731, 0.82794, 0.41194, 0.45445, 0.34951, 0.21288, 1.9364, 3.8839, 2.2123, 0.00028516, 0.00028516, 0.00028516
91
- 90, 0.33818, 0.23887, 0.82712, 0.38309, 0.47201, 0.3563, 0.21723, 1.8942, 3.5157, 2.1923, 0.00028384, 0.00028384, 0.00028384
92
- 91, 0.33817, 0.23916, 0.82471, 0.37115, 0.47856, 0.3677, 0.22534, 1.9206, 3.7495, 2.2067, 0.00028252, 0.00028252, 0.00028252
93
- 92, 0.30746, 0.2204, 0.81511, 0.41337, 0.48853, 0.38436, 0.23793, 1.9158, 3.3119, 2.2045, 0.0002812, 0.0002812, 0.0002812
94
- 93, 0.32205, 0.22693, 0.82014, 0.42726, 0.4339, 0.35789, 0.2161, 1.9268, 3.2917, 2.2452, 0.00027988, 0.00027988, 0.00027988
95
- 94, 0.2972, 0.21774, 0.81642, 0.38111, 0.52191, 0.35675, 0.21237, 1.8947, 3.686, 2.2377, 0.00027856, 0.00027856, 0.00027856
96
- 95, 0.32294, 0.22957, 0.82239, 0.41481, 0.50092, 0.3848, 0.22727, 1.9184, 3.5746, 2.2383, 0.00027724, 0.00027724, 0.00027724
97
- 96, 0.31612, 0.22645, 0.82021, 0.3793, 0.47952, 0.37292, 0.22048, 1.9126, 3.3464, 2.236, 0.00027592, 0.00027592, 0.00027592
98
- 97, 0.31757, 0.22508, 0.81672, 0.36189, 0.46282, 0.35399, 0.20533, 2.0097, 3.8115, 2.399, 0.0002746, 0.0002746, 0.0002746
99
- 98, 0.31277, 0.22137, 0.81856, 0.32361, 0.48973, 0.3481, 0.20856, 2.0474, 3.6429, 2.4017, 0.00027328, 0.00027328, 0.00027328
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
blender_data-29/runs/detect/train/results.png DELETED
Binary file (352 kB)
 
blender_data-29/runs/detect/train/roboflow_deploy.zip DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0692b4591360018389a214a15a09c0b53d836ed485e0521fad75999974c985b1
3
- size 48347844
 
 
 
 
blender_data-29/runs/detect/train/state_dict.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d47286d25dadee20cd04702df0ea71015697a9bbf66f60985e2f6709a6592cc
3
- size 51961561
 
 
 
 
blender_data-29/runs/detect/train/train_batch0.jpg DELETED
Binary file (738 kB)
 
blender_data-29/runs/detect/train/train_batch1.jpg DELETED

Git LFS Details

  • SHA256: 49dea72d7e56df4f8be3a9058f7ad8802ed305ed64bc8b3f1d6501c4542e2482
  • Pointer size: 132 Bytes
  • Size of remote file: 1.01 MB
blender_data-29/runs/detect/train/train_batch2.jpg DELETED
Binary file (751 kB)
 
blender_data-29/runs/detect/train/val_batch0_labels.jpg DELETED
Binary file (754 kB)
 
blender_data-29/runs/detect/train/val_batch0_pred.jpg DELETED
Binary file (876 kB)
 
blender_data-29/runs/detect/train/val_batch1_labels.jpg DELETED
Binary file (755 kB)
 
blender_data-29/runs/detect/train/val_batch1_pred.jpg DELETED
Binary file (818 kB)
 
blender_data-29/runs/detect/train/val_batch2_labels.jpg DELETED
Binary file (765 kB)
 
blender_data-29/runs/detect/train/val_batch2_pred.jpg DELETED
Binary file (854 kB)
 
blender_data-29/runs/detect/train/weights/best.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:087cb0a519020b672f0ba40faca6eec14eae31aed272690ad10b7d6a36581460
3
- size 52098817
 
 
 
 
blender_data-29/runs/detect/train/weights/last.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:07ef21f2842a972013d5f7530c0036e21e9a4e1872c39a008c661d3ab0ed36f1
3
- size 52104769