Object Detection
FBAGSTM commited on
Commit
b4702f8
·
verified ·
1 Parent(s): 773d766

Release AI-ModelZoo-4.0.0

Browse files
Files changed (1) hide show
  1. README.md +15 -31
README.md CHANGED
@@ -5,15 +5,15 @@ license_link: >-
5
  https://github.com/STMicroelectronics/stm32ai-modelzoo/raw/refs/heads/main/object_detection/yolov8n/LICENSE.md
6
  pipeline_tag: object-detection
7
  ---
8
- # Yolov8n object detection quantized
9
 
10
  ## **Use case** : `Object detection`
11
 
12
  # Model description
13
 
14
- Yolov8n is a lightweight and efficient object detection model designed for instance segmentation tasks. It is part of the YOLO (You Only Look Once) family of models, known for their real-time object detection capabilities. The "n" in Yolov8n_seg indicates that it is a nano version, optimized for speed and resource efficiency, making it suitable for deployment on devices with limited computational power, such as mobile devices and embedded systems.
15
 
16
- Yolov8n is implemented in Pytorch by Ultralytics and is quantized in int8 format using tensorflow lite converter.
17
 
18
  ## Network information
19
 
@@ -48,7 +48,7 @@ With an image resolution of NxM and K classes to detect:
48
  | STM32U5 | [] | [] |
49
  | STM32H7 | [] | [] |
50
  | STM32MP1 | [] | [] |
51
- | STM32MP2 | [x] | [x] |
52
  | STM32N6 | [x] | [x] |
53
 
54
 
@@ -56,51 +56,35 @@ With an image resolution of NxM and K classes to detect:
56
 
57
  ## Metrics
58
 
59
- Measures are done with default STM32Cube.AI configuration with enabled input / output allocated option.
60
  > [!CAUTION]
61
- > All YOLOv8 hyperlinks in the tables below link to an external GitHub folder, which is subject to its own license terms:
62
  https://github.com/stm32-hotspot/ultralytics/blob/main/LICENSE
63
  Please also check the folder's README.md file for detailed information about its use and content:
64
  https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/README.md
65
 
66
-
67
  ### Reference **NPU** memory footprint based on COCO Person dataset (see Accuracy for details on dataset)
68
- | Model | Dataset | Format | Resolution | Series | Internal RAM | External RAM | Weights Flash | STM32Cube.AI version | STEdgeAI Core version |
69
- |---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------|----------|--------------|----------|----------------|----------------|-----------------|------------------------|-------------------------|
70
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_192_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 192x192x3 | STM32N6 | 261 | 0 | 2936.52 | 10.2.0 | 2.2.0 |
71
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_256_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 256x256x3 | STM32N6 | 624 | 0 | 2941.09 | 10.2.0 | 2.2.0 |
72
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_320_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 320x320x3 | STM32N6 | 839.06 | 0 | 2947.02 | 10.2.0 | 2.2.0 |
73
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_416_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 416x416x3 | STM32N6 | 2242.84 | 0 | 2958.34 | 10.2.0 | 2.2.0 |
74
  ### Reference **NPU** inference time based on COCO Person dataset (see Accuracy for details on dataset)
75
- | Model | Dataset | Format | Resolution | Board | Execution Engine | Inference time (ms) | Inf / sec | STM32Cube.AI version | STEdgeAI Core version |
76
- |---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------|----------|--------------|---------------|--------------------|-----------------------|-------------|------------------------|-------------------------|
77
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_192_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 192x192x3 | STM32N6570-DK | NPU/MCU | 16.88 | 59.24 | 10.2.0 | 2.2.0 |
78
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_256_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 256x256x3 | STM32N6570-DK | NPU/MCU | 24.94 | 40.1 | 10.2.0 | 2.2.0 |
79
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_320_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 320x320x3 | STM32N6570-DK | NPU/MCU | 31.75 | 31.5 | 10.2.0 | 2.2.0 |
80
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_416_quant_pc_uf_od_coco-person.tflite) | COCO-Person | Int8 | 416x416x3 | STM32N6570-DK | NPU/MCU | 53.11 | 18.83 | 10.2.0 | 2.2.0 |
81
- ### Reference **MPU** inference time based on COCO Person dataset (see Accuracy for details on dataset)
82
- Model | Format | Resolution | Quantization | Board | Execution Engine | Frequency | Inference time (ms) | %NPU | %GPU | %CPU | X-LINUX-AI version | Framework |
83
- |-----------|--------|------------|---------------|-------------------|------------------|-----------|---------------------|-------|-------|------|--------------------|-----------------------|
84
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_256_quant_pc_uf_pose_coco-st.tflite) | Int8 | 256x256x3 | per-channel** | STM32MP257F-DK2 | NPU/GPU | 800 MHz | 102.8 ms | 11.70 | 88.30 |0 | v6.1.0 | OpenVX |
85
- | [YOLOv8n per tensor](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_256_quant_pt_uf_pose_coco-st.tflite) | Int8 | 256x256x3 | per-tensor | STM32MP257F-DK2 | NPU/GPU | 800 MHz | 17.57 ms | 86.79 | 13.21 |0 | v6.1.0 | OpenVX |
86
-
87
- ** **To get the most out of MP25 NPU hardware acceleration, please use per-tensor quantization**
88
 
89
  ### AP on COCO Person dataset
90
 
91
-
92
  Dataset details: [link](https://cocodataset.org/#download) , License [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/legalcode) , Quotation[[1]](#1) , Number of classes: 80, Number of images: 118,287
93
 
94
 
95
  | Model | Format | Resolution | AP* |
96
  |-------|--------|------------|----------------|
97
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_192_quant_pc_uf_od_coco-person.tflite) | Int8 | 192x192x3 | 53.50 % |
98
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_256_quant_pc_uf_od_coco-person.tflite) | Int8 | 256x256x3 | 58.40 % |
99
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_320_quant_pc_uf_od_coco-person.tflite) | Int8 | 320x320x3 | 61.80 % |
100
- | [YOLOv8n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolov8n_416_quant_pc_uf_od_coco-person.tflite) | Int8 | 416x416x3 | 64.80 % |
101
 
102
  \* EVAL_IOU = 0.5, NMS_THRESH = 0.5, SCORE_THRESH = 0.001, MAX_DETECTIONS = 100
103
 
 
104
  ## Integration in a simple example and other services support:
105
 
106
  Please refer to the stm32ai-modelzoo-services GitHub [here](https://github.com/STMicroelectronics/stm32ai-modelzoo-services).
 
5
  https://github.com/STMicroelectronics/stm32ai-modelzoo/raw/refs/heads/main/object_detection/yolov8n/LICENSE.md
6
  pipeline_tag: object-detection
7
  ---
8
+ # Yolo11n object detection quantized
9
 
10
  ## **Use case** : `Object detection`
11
 
12
  # Model description
13
 
14
+ Yolo11n is a lightweight and efficient object detection model designed for instance segmentation tasks. It is part of the YOLO (You Only Look Once) family of models, known for their real-time object detection capabilities. The "n" in Yolo11n indicates that it is a nano version, optimized for speed and resource efficiency, making it suitable for deployment on devices with limited computational power, such as mobile devices and embedded systems.
15
 
16
+ Yolo11n is implemented in Pytorch by Ultralytics and is quantized in int8 format using tensorflow lite converter.
17
 
18
  ## Network information
19
 
 
48
  | STM32U5 | [] | [] |
49
  | STM32H7 | [] | [] |
50
  | STM32MP1 | [] | [] |
51
+ | STM32MP2 | [] | [] |
52
  | STM32N6 | [x] | [x] |
53
 
54
 
 
56
 
57
  ## Metrics
58
 
59
+ Measures are done with default STM32Cube.AI configuration with enabled input / output allocated option.
60
  > [!CAUTION]
61
+ > All YOLOv11 hyperlinks in the tables below link to an external GitHub folder, which is subject to its own license terms:
62
  https://github.com/stm32-hotspot/ultralytics/blob/main/LICENSE
63
  Please also check the folder's README.md file for detailed information about its use and content:
64
  https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/README.md
65
 
 
66
  ### Reference **NPU** memory footprint based on COCO Person dataset (see Accuracy for details on dataset)
67
+ | Model | Dataset | Format | Resolution | Series | Internal RAM | External RAM | Weights Flash | STEdgeAI Core version |
68
+ |-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------|----------|--------------|----------|----------------|----------------|-----------------|-------------------------|
69
+ | [YOLO11n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolo11/yolo11n_256_quant_pc_uf_od_coco-person-st.tflite) | COCO-Person | Int8 | 256x256x3 | STM32N6 | 656 | 0 | 2535.83 | 3.0.0 |
70
+
 
 
71
  ### Reference **NPU** inference time based on COCO Person dataset (see Accuracy for details on dataset)
72
+ | Model | Dataset | Format | Resolution | Board | Execution Engine | Inference time (ms) | Inf / sec | STEdgeAI Core version |
73
+ |-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------|----------|--------------|---------------|--------------------|-----------------------|-------------|-------------------------|
74
+ | [YOLO11n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolo11/yolo11n_256_quant_pc_uf_od_coco-person-st.tflite) | COCO-Person | Int8 | 256x256x3 | STM32N6570-DK | NPU/MCU | 26.37 | 36.50 | 3.0.0 |
 
 
 
 
 
 
 
 
 
 
75
 
76
  ### AP on COCO Person dataset
77
 
 
78
  Dataset details: [link](https://cocodataset.org/#download) , License [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/legalcode) , Quotation[[1]](#1) , Number of classes: 80, Number of images: 118,287
79
 
80
 
81
  | Model | Format | Resolution | AP* |
82
  |-------|--------|------------|----------------|
83
+ | [YOLOv11n per channel](https://github.com/stm32-hotspot/ultralytics/blob/main/examples/YOLOv8-STEdgeAI/stedgeai_models/object_detection/yolo11/yolo11n_256_quant_pc_uf_od_coco-person-st.tflite) | Int8 | 640x640x3 | 64 % |
 
 
 
84
 
85
  \* EVAL_IOU = 0.5, NMS_THRESH = 0.5, SCORE_THRESH = 0.001, MAX_DETECTIONS = 100
86
 
87
+
88
  ## Integration in a simple example and other services support:
89
 
90
  Please refer to the stm32ai-modelzoo-services GitHub [here](https://github.com/STMicroelectronics/stm32ai-modelzoo-services).