File size: 11,141 Bytes
667c9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f0b6d53
b536c0c
41045d4
667c9fe
b536c0c
 
667c9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b536c0c
667c9fe
 
 
 
 
cc5fcca
 
667c9fe
cc5fcca
 
667c9fe
 
 
 
b536c0c
667c9fe
 
 
 
cc5fcca
667c9fe
cc5fcca
 
 
 
 
 
b536c0c
667c9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc5fcca
667c9fe
 
 
 
 
 
 
 
 
cc5fcca
667c9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b536c0c
667c9fe
 
 
 
 
 
 
 
 
cc5fcca
667c9fe
 
 
 
 
 
 
 
 
 
b536c0c
667c9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b536c0c
667c9fe
 
 
 
 
b536c0c
 
 
 
 
 
 
667c9fe
 
 
 
 
f0b6d53
667c9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
---
license: apache-2.0
library_name: edgefirst
pipeline_tag: object-detection
tags:
  - edge-ai
  - npu
  - tflite
  - onnx
  - int8
  - yolo
  - gstreamer
  - edgefirst
  - nxp
  - hailo
  - jetson
  - real-time
  - embedded
  - multiplatform
model-index:
  - name: yolo11-det
    results:
      - task:
          type: object-detection
        dataset:
          name: COCO val2017
          type: coco
        metrics:
          - name: "mAP@0.5 (Nano ONNX FP32)"
            type: map_50
            value: 53.4
          - name: "mAP@0.5-0.95 (Nano ONNX FP32)"
            type: map
            value: 37.9
          - name: "mAP@0.5 (Nano TFLite INT8)"
            type: map_50
            value: 50.1
          - name: "mAP@0.5-0.95 (Nano TFLite INT8)"
            type: map
            value: 34.5
---

# YOLO11 Detection β€” EdgeFirst Edge AI

**NXP i.MX 8M Plus** | **NXP i.MX 93** | **NXP i.MX 95** | **NXP Ara240** | **RPi5 + Hailo-8/8L** | **NVIDIA Jetson**
YOLO11 Detection models optimized for edge AI deployment across multiple hardware platforms. All sizes from Nano to XLarge, in ONNX FP32 and TFLite INT8 formats, with platform-specific compiled models for NPU acceleration.

Trained on [COCO 2017](https://test.edgefirst.studio/public/projects/2839/home) (80 classes). Part of the [EdgeFirst Model Zoo](https://huggingface.co/spaces/EdgeFirst/Models).
> [!TIP]
> **Training session**: [View on EdgeFirst Studio](https://test.edgefirst.studio/public/projects/2839/experiment/training/list?exp_id=4656) β€” dataset, training config, metrics, and exported artifacts.

> [!NOTE]
> Newer architecture with attention blocks.

---

## Size Comparison

All models validated on COCO val2017 (5000 images, 80 classes).

| Size | Params | GFLOPs | ONNX FP32 mAP@0.5 | ONNX FP32 mAP@0.5-0.95 | TFLite INT8 mAP@0.5 | TFLite INT8 mAP@0.5-0.95 |
|------|--------|--------|--------------------|-------------------------|----------------------|--------------------------|
| Nano | 2.6M | 6.5 | 53.4% | 37.9% | 50.1% | 34.5% |
| Small | 9.4M | 21.5 | β€” | β€” | β€” | β€” |
| Medium | 20.1M | 68.0 | β€” | β€” | β€” | β€” |
| Large | 25.3M | 87.6 | β€” | β€” | β€” | β€” |
| XLarge | 56.9M | 195.0 | β€” | β€” | β€” | β€” |

---

## On-Target Performance

Full pipeline timing: pre-processing + inference + post-processing.

| Size | Platform | Pre-proc (ms) | Inference (ms) | Post-proc (ms) | Total (ms) | FPS |
|------|----------|---------------|----------------|-----------------|------------|-----|
| β€” | β€” | β€” | β€” | β€” | β€” | β€” |

*Measured with [EdgeFirst Perception](https://github.com/EdgeFirstAI) stack. Timing includes full GStreamer pipeline overhead.*

---

## Downloads

<details open>
<summary><strong>ONNX FP32</strong> β€” Any platform with ONNX Runtime.</summary>

| Size | File | Status |
|------|------|--------|
| Nano | `yolo11n-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11n-det-coco.onnx) |
| Small | `yolo11s-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11s-det-coco.onnx) |
| Medium | `yolo11m-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11m-det-coco.onnx) |
| Large | `yolo11l-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11l-det-coco.onnx) |
| XLarge | `yolo11x-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11x-det-coco.onnx) |

</details>

<details>
<summary><strong>TFLite INT8</strong> β€” CPU or NPU via runtime delegate (i.MX 8M Plus VX Delegate).</summary>

| Size | File | Status |
|------|------|--------|
| Nano | `yolo11n-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11n-det-coco.tflite) |
| Small | `yolo11s-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11s-det-coco.tflite) |
| Medium | `yolo11m-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11m-det-coco.tflite) |
| Large | `yolo11l-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11l-det-coco.tflite) |
| XLarge | `yolo11x-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11x-det-coco.tflite) |

</details>



---

## Deploy with EdgeFirst Perception

Copy-paste [GStreamer](https://github.com/EdgeFirstAI/gstreamer) pipeline examples for each platform.

### NXP i.MX 8M Plus β€” Camera to Detection with Vivante NPU

```bash
gst-launch-1.0 \
  v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
  edgefirstcameraadaptor ! \
  tensor_filter framework=tensorflow-lite \
    model=yolo11n-det-coco.tflite \
    custom=Delegate:External,ExtDelegateLib:libvx_delegate.so ! \
  edgefirstdetdecoder ! edgefirstoverlay ! waylandsink
```

### RPi5 + Hailo-8L

```bash
gst-launch-1.0 \
  v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
  hailonet hef-path=yolo11n-det-coco.hailo8l.hef ! \
  hailofilter function-name=yolo11_nms ! \
  hailooverlay ! videoconvert ! autovideosink
```

### NVIDIA Jetson (TensorRT)

```bash
gst-launch-1.0 \
  v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
  edgefirstcameraadaptor ! \
  nvinfer config-file-path=yolo11n-det-coco-config.txt ! \
  edgefirstdetdecoder ! edgefirstoverlay ! nveglglessink
```


*Full pipeline documentation: [EdgeFirst GStreamer Plugins](https://github.com/EdgeFirstAI/gstreamer)*

---

## Foundation (HAL) Python Integration

```python
from edgefirst.hal import Model, TensorImage

# Load model β€” metadata (labels, decoder config) is embedded in the file
model = Model("yolo11n-det-coco.tflite")

# Run inference on an image
image = TensorImage.from_file("image.jpg")
results = model.predict(image)

# Access detections
for det in results.detections:
    print(f"{det.label}: {det.confidence:.2f} at {det.bbox}")
```

*[EdgeFirst HAL](https://github.com/EdgeFirstAI/hal) β€” Hardware abstraction layer with accelerated inference delegates.*

---

## CameraAdaptor

EdgeFirst [CameraAdaptor](https://github.com/EdgeFirstAI/cameraadaptor) enables training and inference directly on native sensor formats (GREY, YUYV, etc.) β€” skipping the ISP color conversion pipeline entirely. This reduces latency and power consumption on edge devices.

CameraAdaptor variants are included alongside baseline RGB models:

| Variant | Input Format | Use Case |
|---------|-------------|----------|
| `yolo11n-det-coco.onnx` | RGB (3ch) | Standard camera input |
| `yolo11n-det-coco-grey.onnx` | GREY (1ch) | Monochrome / IR sensors |
| `yolo11n-det-coco-yuyv.onnx` | YUYV (2ch) | Raw sensor bypass |

*Train CameraAdaptor models with [EdgeFirst Studio](https://edgefirst.studio) β€” the CameraAdaptor layer is automatically inserted during training.*

---

## Train Your Own with EdgeFirst Studio

Train on your own dataset with [**EdgeFirst Studio**](https://edgefirst.studio):

- **Free tier** includes YOLO training with automatic INT8 quantization and edge deployment
- Upload datasets via [EdgeFirst Recorder](https://github.com/EdgeFirstAI/recorder) or COCO/YOLO format
- AI-assisted annotation with auto-labeling
- CameraAdaptor integration for native sensor format training
- Deploy trained models to edge devices via [EdgeFirst Client](https://github.com/EdgeFirstAI/client)

---

## See Also

Other models in the [EdgeFirst Model Zoo](https://huggingface.co/spaces/EdgeFirst/Models):

| Model | Task | Best Nano Metric | Link |
|-------|------|-------------------|------|
| YOLOv5 Detection | Detection | 49.6% mAP@0.5 (ONNX) | [EdgeFirst/yolov5-det](https://huggingface.co/EdgeFirst/yolov5-det) |
| YOLOv8 Detection | Detection | 50.2% mAP@0.5 (ONNX) | [EdgeFirst/yolov8-det](https://huggingface.co/EdgeFirst/yolov8-det) |
| YOLOv8 Segmentation | Segmentation | 34.1% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolov8-seg](https://huggingface.co/EdgeFirst/yolov8-seg) |
| YOLO11 Segmentation | Segmentation | 35.5% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolo11-seg](https://huggingface.co/EdgeFirst/yolo11-seg) |
| YOLO26 Detection | Detection | 54.9% mAP@0.5 (ONNX) | [EdgeFirst/yolo26-det](https://huggingface.co/EdgeFirst/yolo26-det) |
| YOLO26 Segmentation | Segmentation | 37.0% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolo26-seg](https://huggingface.co/EdgeFirst/yolo26-seg) |

---

## Technical Details

### Quantization Pipeline

All TFLite INT8 models are produced by EdgeFirst's custom quantization pipeline ([details](https://github.com/EdgeFirstAI/studio-ultralytics)):

1. **ONNX Export** β€” Standard Ultralytics export with `simplify=True`
2. **TF-Wrapped ONNX** β€” Box coordinates normalized to [0,1] inside DFL decode via `tf_wrapper` (~1.2% better mAP than post-hoc normalization)
3. **Split Decoder** β€” Boxes, scores, and mask coefficients split into separate output tensors for independent INT8 quantization scales
4. **Smart Calibration** β€” 500 images selected via greedy coverage maximization from COCO val2017
5. **Full INT8** β€” `uint8` input (raw pixels), `int8` output (per-tensor scales), MLIR quantizer

### Split Decoder Output Format

**Detection** (e.g., yolo11n):
- Boxes: `(1, 4, 8400)` β€” normalized [0,1] coordinates
- Scores: `(1, 80, 8400)` β€” class probabilities

Each tensor has independent quantization scale and zero-point. EdgeFirst HAL handles dequantization and reassembly automatically.

### Metadata

- **TFLite**: `edgefirst.json`, `labels.txt`, and `edgefirst.yaml` embedded via ZIP (no `tflite-support` dependency)
- **ONNX**: `edgefirst.json` embedded via `model.metadata_props`

No standalone metadata files β€” models are self-contained.

---

## Limitations

- **COCO bias** β€” Models trained on COCO (80 classes) inherit its biases: Western-centric scenes, specific object distributions, limited weather/lighting diversity
- **INT8 accuracy loss** β€” Full-integer quantization typically degrades mAP by 6-12% relative to FP32; actual loss depends on model architecture and dataset
- **Thermal variation** β€” On-target performance varies with device temperature; sustained inference may throttle on passively-cooled devices
- **Input resolution** β€” All models expect 640Γ—640 input; other resolutions require letterboxing or may reduce accuracy
- **CameraAdaptor variants** β€” GREY/YUYV models trade color information for latency; accuracy may differ from RGB baseline depending on the task

---

## Citation

```bibtex
@software{edgefirst_yolo11_det,
  title = { {YOLO11 Detection β€” EdgeFirst Edge AI} },
  author = {Au-Zone Technologies},
  url = {https://huggingface.co/EdgeFirst/yolo11-det},
  year = {2026},
  license = {Apache-2.0},
}
```

---

<p align="center">
  <sub>
    <a href="https://edgefirst.studio">EdgeFirst Studio</a> Β· <a href="https://github.com/EdgeFirstAI">GitHub</a> Β· <a href="https://doc.edgefirst.ai">Docs</a> Β· <a href="https://www.au-zone.com">Au-Zone Technologies</a><br>
    Apache 2.0 Β· Β© Au-Zone Technologies Inc.
  </sub>
</p>