sebastientaylor commited on
Commit
667c9fe
Β·
verified Β·
1 Parent(s): 53bac00

Update model card for yolo11-det

Browse files
Files changed (1) hide show
  1. README.md +348 -0
README.md ADDED
@@ -0,0 +1,348 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: edgefirst
4
+ pipeline_tag: object-detection
5
+ tags:
6
+ - edge-ai
7
+ - npu
8
+ - tflite
9
+ - onnx
10
+ - int8
11
+ - yolo
12
+ - gstreamer
13
+ - edgefirst
14
+ - nxp
15
+ - hailo
16
+ - jetson
17
+ - real-time
18
+ - embedded
19
+ - multiplatform
20
+ model-index:
21
+ - name: yolo11-det
22
+ results:
23
+ - task:
24
+ type: object-detection
25
+ dataset:
26
+ name: COCO val2017
27
+ type: coco
28
+ metrics:
29
+ - name: "mAP@0.5 (Nano ONNX FP32)"
30
+ type: map_50
31
+ value: 53.4
32
+ - name: "mAP@0.5-0.95 (Nano ONNX FP32)"
33
+ type: map
34
+ value: 37.9
35
+ - name: "mAP@0.5 (Nano TFLite INT8)"
36
+ type: map_50
37
+ value: 50.1
38
+ - name: "mAP@0.5-0.95 (Nano TFLite INT8)"
39
+ type: map
40
+ value: 34.5
41
+ ---
42
+
43
+ # YOLO11 Detection β€” EdgeFirst Edge AI
44
+
45
+ <p align="center">
46
+ <img src="assets/demo.jpg" alt="YOLO11 Detection demo" width="640">
47
+ </p>
48
+
49
+ **NXP i.MX 8M Plus** | **NXP i.MX 93** | **NXP i.MX 95** | **NXP Ara240** | **RPi5 + Hailo-8/8L** | **NVIDIA Jetson**
50
+ YOLO11 Detection models optimized for edge AI deployment across multiple hardware platforms. All sizes from Nano to XLarge, in ONNX FP32 and TFLite INT8 formats, with platform-specific compiled models for NPU acceleration.
51
+
52
+ Trained on [COCO 2017](https://test.edgefirst.studio/public/projects/1123/datasets/gallery/main?dataset=4819) (80 classes). Part of the [EdgeFirst Model Zoo](https://huggingface.co/EdgeFirst).
53
+ > **Training session**: [View on EdgeFirst Studio](https://test.edgefirst.studio/public/projects/1123/experiment/training/details?train_session_id=9506) β€” dataset, training config, metrics, and exported artifacts.
54
+
55
+ > **Note**: Newer architecture with attention blocks.
56
+
57
+ ---
58
+
59
+ ## Size Comparison
60
+
61
+ All models validated on COCO val2017 (5000 images, 80 classes).
62
+
63
+ | Size | Params | GFLOPs | ONNX FP32 mAP@0.5 | ONNX FP32 mAP@0.5-0.95 | TFLite INT8 mAP@0.5 | TFLite INT8 mAP@0.5-0.95 |
64
+ |------|--------|--------|--------------------|-------------------------|----------------------|--------------------------|
65
+ | Nano | 2.6M | 6.5 | 53.4% | 37.9% | 50.1% | 34.5% |
66
+ | Small | 9.4M | 21.5 | β€” | β€” | β€” | β€” |
67
+ | Medium | 20.1M | 68.0 | β€” | β€” | β€” | β€” |
68
+ | Large | 25.3M | 87.6 | β€” | β€” | β€” | β€” |
69
+ | XLarge | 56.9M | 195.0 | β€” | β€” | β€” | β€” |
70
+
71
+ ---
72
+
73
+ ## On-Target Performance
74
+
75
+ Full pipeline timing: pre-processing + inference + post-processing.
76
+
77
+ | Size | Platform | Pre-proc (ms) | Inference (ms) | Post-proc (ms) | Total (ms) | FPS |
78
+ |------|----------|---------------|----------------|-----------------|------------|-----|
79
+ | β€” | β€” | β€” | β€” | β€” | β€” | β€” |
80
+
81
+ > Measured with [EdgeFirst Perception](https://github.com/EdgeFirstAI) stack. Timing includes full GStreamer pipeline overhead.
82
+
83
+ ---
84
+
85
+ ## Downloads
86
+
87
+ ### Universal Formats
88
+
89
+ <details>
90
+ <summary><strong>ONNX FP32</strong> β€” Any platform with ONNX Runtime</summary>
91
+
92
+ | Size | File | Download |
93
+ |------|------|----------|
94
+ | Nano | `yolo11n-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11n-det-coco.onnx) |
95
+ | Small | `yolo11s-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11s-det-coco.onnx) |
96
+ | Medium | `yolo11m-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11m-det-coco.onnx) |
97
+ | Large | `yolo11l-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11l-det-coco.onnx) |
98
+ | XLarge | `yolo11x-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11x-det-coco.onnx) |
99
+
100
+ </details>
101
+
102
+ <details>
103
+ <summary><strong>TFLite INT8</strong> β€” Any platform with TFLite (i.MX 8M Plus uses VX delegate)</summary>
104
+
105
+ | Size | File | Download |
106
+ |------|------|----------|
107
+ | Nano | `yolo11n-det-coco-int8.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11n-det-coco-int8.tflite) |
108
+ | Small | `yolo11s-det-coco-int8.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11s-det-coco-int8.tflite) |
109
+ | Medium | `yolo11m-det-coco-int8.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11m-det-coco-int8.tflite) |
110
+ | Large | `yolo11l-det-coco-int8.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11l-det-coco-int8.tflite) |
111
+ | XLarge | `yolo11x-det-coco-int8.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11x-det-coco-int8.tflite) |
112
+
113
+ </details>
114
+
115
+ ### Platform-Specific
116
+
117
+ <details>
118
+ <summary><strong>NXP i.MX 93</strong> β€” Ethos-U NPU via ARM VELA compiler.</summary>
119
+
120
+ | Size | File | Download |
121
+ |------|------|----------|
122
+ | Nano | `yolo11n-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx93/yolo11n-det-coco.tflite) |
123
+ | Small | `yolo11s-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx93/yolo11s-det-coco.tflite) |
124
+ | Medium | `yolo11m-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx93/yolo11m-det-coco.tflite) |
125
+ | Large | `yolo11l-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx93/yolo11l-det-coco.tflite) |
126
+ | XLarge | `yolo11x-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx93/yolo11x-det-coco.tflite) |
127
+
128
+ </details>
129
+
130
+ <details>
131
+ <summary><strong>NXP i.MX 95</strong> β€” eIQ Neutron NPU optimized.</summary>
132
+
133
+ | Size | File | Download |
134
+ |------|------|----------|
135
+ | Nano | `yolo11n-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx95/yolo11n-det-coco.tflite) |
136
+ | Small | `yolo11s-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx95/yolo11s-det-coco.tflite) |
137
+ | Medium | `yolo11m-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx95/yolo11m-det-coco.tflite) |
138
+ | Large | `yolo11l-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx95/yolo11l-det-coco.tflite) |
139
+ | XLarge | `yolo11x-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/imx95/yolo11x-det-coco.tflite) |
140
+
141
+ </details>
142
+
143
+ <details>
144
+ <summary><strong>NXP Ara240</strong> β€” Kinara DVM compiled model.</summary>
145
+
146
+ | Size | File | Download |
147
+ |------|------|----------|
148
+ | Nano | `yolo11n-det-coco.dvm` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/ara240/yolo11n-det-coco.dvm) |
149
+ | Small | `yolo11s-det-coco.dvm` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/ara240/yolo11s-det-coco.dvm) |
150
+ | Medium | `yolo11m-det-coco.dvm` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/ara240/yolo11m-det-coco.dvm) |
151
+ | Large | `yolo11l-det-coco.dvm` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/ara240/yolo11l-det-coco.dvm) |
152
+ | XLarge | `yolo11x-det-coco.dvm` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/ara240/yolo11x-det-coco.dvm) |
153
+
154
+ </details>
155
+
156
+ <details>
157
+ <summary><strong>RPi5 + Hailo-8/8L</strong> β€” Hailo-8L (13 TOPS) and Hailo-8 (26 TOPS).</summary>
158
+
159
+ | Size | File | Download |
160
+ |------|------|----------|
161
+ | Nano | `yolo11n-det-coco.hef` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/hailo/yolo11n-det-coco.hef) |
162
+ | Small | `yolo11s-det-coco.hef` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/hailo/yolo11s-det-coco.hef) |
163
+ | Medium | `yolo11m-det-coco.hef` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/hailo/yolo11m-det-coco.hef) |
164
+ | Large | `yolo11l-det-coco.hef` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/hailo/yolo11l-det-coco.hef) |
165
+ | XLarge | `yolo11x-det-coco.hef` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/hailo/yolo11x-det-coco.hef) |
166
+
167
+ </details>
168
+
169
+ <details>
170
+ <summary><strong>NVIDIA Jetson</strong> β€” Jetson FP16 and INT8 engines.</summary>
171
+
172
+ | Size | File | Download |
173
+ |------|------|----------|
174
+ | Nano | `yolo11n-det-coco.engine` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/jetson/yolo11n-det-coco.engine) |
175
+ | Small | `yolo11s-det-coco.engine` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/jetson/yolo11s-det-coco.engine) |
176
+ | Medium | `yolo11m-det-coco.engine` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/jetson/yolo11m-det-coco.engine) |
177
+ | Large | `yolo11l-det-coco.engine` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/jetson/yolo11l-det-coco.engine) |
178
+ | XLarge | `yolo11x-det-coco.engine` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/jetson/yolo11x-det-coco.engine) |
179
+
180
+ </details>
181
+
182
+
183
+ ---
184
+
185
+ ## Deploy with EdgeFirst Perception
186
+
187
+ Copy-paste [GStreamer](https://github.com/EdgeFirstAI/gstreamer) pipeline examples for each platform.
188
+
189
+ ### NXP i.MX 8M Plus β€” Camera to Detection with Vivante NPU
190
+
191
+ ```bash
192
+ gst-launch-1.0 \
193
+ v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
194
+ edgefirstcameraadaptor ! \
195
+ tensor_filter framework=tensorflow-lite \
196
+ model=yolo11n-det-coco-int8.tflite \
197
+ custom=Delegate:External,ExtDelegateLib:libvx_delegate.so ! \
198
+ edgefirstdetdecoder ! edgefirstoverlay ! waylandsink
199
+ ```
200
+
201
+ ### RPi5 + Hailo-8L
202
+
203
+ ```bash
204
+ gst-launch-1.0 \
205
+ v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
206
+ hailonet hef-path=yolo11n-det-coco-h8l.hef ! \
207
+ hailofilter function-name=yolo11_nms ! \
208
+ hailooverlay ! videoconvert ! autovideosink
209
+ ```
210
+
211
+ ### NVIDIA Jetson (TensorRT)
212
+
213
+ ```bash
214
+ gst-launch-1.0 \
215
+ v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
216
+ edgefirstcameraadaptor ! \
217
+ nvinfer config-file-path=yolo11n-det-coco-config.txt ! \
218
+ edgefirstdetdecoder ! edgefirstoverlay ! nveglglessink
219
+ ```
220
+
221
+
222
+ > Full pipeline documentation: [EdgeFirst GStreamer Plugins](https://github.com/EdgeFirstAI/gstreamer)
223
+
224
+ ---
225
+
226
+ ## Foundation (HAL) Python Integration
227
+
228
+ ```python
229
+ from edgefirst.hal import Model, TensorImage
230
+
231
+ # Load model β€” metadata (labels, decoder config) is embedded in the file
232
+ model = Model("yolo11n-det-coco-int8.tflite")
233
+
234
+ # Run inference on an image
235
+ image = TensorImage.from_file("image.jpg")
236
+ results = model.predict(image)
237
+
238
+ # Access detections
239
+ for det in results.detections:
240
+ print(f"{det.label}: {det.confidence:.2f} at {det.bbox}")
241
+ ```
242
+
243
+ > [EdgeFirst HAL](https://github.com/EdgeFirstAI/hal) β€” Hardware abstraction layer with accelerated inference delegates.
244
+
245
+ ---
246
+
247
+ ## CameraAdaptor
248
+
249
+ EdgeFirst [CameraAdaptor](https://github.com/EdgeFirstAI/cameraadaptor) enables training and inference directly on native sensor formats (GREY, YUYV, etc.) β€” skipping the ISP color conversion pipeline entirely. This reduces latency and power consumption on edge devices.
250
+
251
+ CameraAdaptor variants are included alongside baseline RGB models:
252
+
253
+ | Variant | Input Format | Use Case |
254
+ |---------|-------------|----------|
255
+ | `yolo11n-det-coco.onnx` | RGB (3ch) | Standard camera input |
256
+ | `yolo11n-det-coco-grey.onnx` | GREY (1ch) | Monochrome / IR sensors |
257
+ | `yolo11n-det-coco-yuyv.onnx` | YUYV (2ch) | Raw sensor bypass |
258
+
259
+ > Train CameraAdaptor models with [EdgeFirst Studio](https://edgefirst.studio) β€” the CameraAdaptor layer is automatically inserted during training.
260
+
261
+ ---
262
+
263
+ ## Train Your Own with EdgeFirst Studio
264
+
265
+ > Train on your own dataset with [**EdgeFirst Studio**](https://edgefirst.studio).
266
+ >
267
+ > - **Free tier** includes YOLO training with automatic INT8 quantization and edge deployment
268
+ > - Upload datasets via [EdgeFirst Recorder](https://github.com/EdgeFirstAI/recorder) or COCO/YOLO format
269
+ > - AI-assisted annotation with auto-labeling
270
+ > - CameraAdaptor integration for native sensor format training
271
+ > - One-click deployment to edge devices via [EdgeFirst Client](https://github.com/EdgeFirstAI/client)
272
+
273
+ ---
274
+
275
+ ## See Also
276
+
277
+ Other models in the [EdgeFirst Model Zoo](https://huggingface.co/EdgeFirst):
278
+
279
+ | Model | Task | Best Nano Metric | Link |
280
+ |-------|------|-------------------|------|
281
+ | YOLOv5 Detection | Detection | 49.6% mAP@0.5 (ONNX) | [EdgeFirst/yolov5-det](https://huggingface.co/EdgeFirst/yolov5-det) |
282
+ | YOLOv8 Detection | Detection | 50.2% mAP@0.5 (ONNX) | [EdgeFirst/yolov8-det](https://huggingface.co/EdgeFirst/yolov8-det) |
283
+ | YOLOv8 Segmentation | Segmentation | 34.1% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolov8-seg](https://huggingface.co/EdgeFirst/yolov8-seg) |
284
+ | YOLO11 Segmentation | Segmentation | 35.5% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolo11-seg](https://huggingface.co/EdgeFirst/yolo11-seg) |
285
+ | YOLO26 Detection | Detection | 54.9% mAP@0.5 (ONNX) | [EdgeFirst/yolo26-det](https://huggingface.co/EdgeFirst/yolo26-det) |
286
+ | YOLO26 Segmentation | Segmentation | 37.0% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolo26-seg](https://huggingface.co/EdgeFirst/yolo26-seg) |
287
+
288
+ ---
289
+
290
+ ## Technical Details
291
+
292
+ ### Quantization Pipeline
293
+
294
+ All TFLite INT8 models are produced by EdgeFirst's custom quantization pipeline ([details](https://github.com/EdgeFirstAI/studio-ultralytics)):
295
+
296
+ 1. **ONNX Export** β€” Standard Ultralytics export with `simplify=True`
297
+ 2. **TF-Wrapped ONNX** β€” Box coordinates normalized to [0,1] inside DFL decode via `tf_wrapper` (~1.2% better mAP than post-hoc normalization)
298
+ 3. **Split Decoder** β€” Boxes, scores, and mask coefficients split into separate output tensors for independent INT8 quantization scales
299
+ 4. **Smart Calibration** β€” 500 images selected via greedy coverage maximization from COCO val2017
300
+ 5. **Full INT8** β€” `uint8` input (raw pixels), `int8` output (per-tensor scales), MLIR quantizer
301
+
302
+ ### Split Decoder Output Format
303
+
304
+ **Detection** (e.g., yolo11n):
305
+ - Boxes: `(1, 4, 8400)` β€” normalized [0,1] coordinates
306
+ - Scores: `(1, 80, 8400)` β€” class probabilities
307
+
308
+ Each tensor has independent quantization scale and zero-point. EdgeFirst HAL handles dequantization and reassembly automatically.
309
+
310
+ ### Metadata
311
+
312
+ - **TFLite**: `edgefirst.json`, `labels.txt`, and `edgefirst.yaml` embedded via ZIP (no `tflite-support` dependency)
313
+ - **ONNX**: `edgefirst.json` embedded via `model.metadata_props`
314
+
315
+ No standalone metadata files β€” models are self-contained.
316
+
317
+ ---
318
+
319
+ ## Limitations
320
+
321
+ - **COCO bias** β€” Models trained on COCO (80 classes) inherit its biases: Western-centric scenes, specific object distributions, limited weather/lighting diversity
322
+ - **INT8 accuracy loss** β€” Full-integer quantization typically degrades mAP by 6-12% relative to FP32; actual loss depends on model architecture and dataset
323
+ - **Thermal variation** β€” On-target performance varies with device temperature; sustained inference may throttle on passively-cooled devices
324
+ - **Input resolution** β€” All models expect 640Γ—640 input; other resolutions require letterboxing or may reduce accuracy
325
+ - **CameraAdaptor variants** β€” GREY/YUYV models trade color information for latency; accuracy may differ from RGB baseline depending on the task
326
+
327
+ ---
328
+
329
+ ## Citation
330
+
331
+ ```bibtex
332
+ @software{edgefirst_yolo11_det,
333
+ title = { {YOLO11 Detection β€” EdgeFirst Edge AI} },
334
+ author = {Au-Zone Technologies},
335
+ url = {https://huggingface.co/EdgeFirst/yolo11-det},
336
+ year = {2026},
337
+ license = {Apache-2.0},
338
+ }
339
+ ```
340
+
341
+ ---
342
+
343
+ <p align="center">
344
+ <sub>
345
+ <a href="https://edgefirst.studio">EdgeFirst Studio</a> Β· <a href="https://github.com/EdgeFirstAI">GitHub</a> Β· <a href="https://doc.edgefirst.ai">Docs</a> Β· <a href="https://www.au-zone.com">Au-Zone Technologies</a><br>
346
+ Apache 2.0 Β· Β© Au-Zone Technologies Inc.
347
+ </sub>
348
+ </p>