File size: 11,141 Bytes
667c9fe f0b6d53 b536c0c 41045d4 667c9fe b536c0c 667c9fe b536c0c 667c9fe cc5fcca 667c9fe cc5fcca 667c9fe b536c0c 667c9fe cc5fcca 667c9fe cc5fcca b536c0c 667c9fe cc5fcca 667c9fe cc5fcca 667c9fe b536c0c 667c9fe cc5fcca 667c9fe b536c0c 667c9fe b536c0c 667c9fe b536c0c 667c9fe f0b6d53 667c9fe | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 | ---
license: apache-2.0
library_name: edgefirst
pipeline_tag: object-detection
tags:
- edge-ai
- npu
- tflite
- onnx
- int8
- yolo
- gstreamer
- edgefirst
- nxp
- hailo
- jetson
- real-time
- embedded
- multiplatform
model-index:
- name: yolo11-det
results:
- task:
type: object-detection
dataset:
name: COCO val2017
type: coco
metrics:
- name: "mAP@0.5 (Nano ONNX FP32)"
type: map_50
value: 53.4
- name: "mAP@0.5-0.95 (Nano ONNX FP32)"
type: map
value: 37.9
- name: "mAP@0.5 (Nano TFLite INT8)"
type: map_50
value: 50.1
- name: "mAP@0.5-0.95 (Nano TFLite INT8)"
type: map
value: 34.5
---
# YOLO11 Detection β EdgeFirst Edge AI
**NXP i.MX 8M Plus** | **NXP i.MX 93** | **NXP i.MX 95** | **NXP Ara240** | **RPi5 + Hailo-8/8L** | **NVIDIA Jetson**
YOLO11 Detection models optimized for edge AI deployment across multiple hardware platforms. All sizes from Nano to XLarge, in ONNX FP32 and TFLite INT8 formats, with platform-specific compiled models for NPU acceleration.
Trained on [COCO 2017](https://test.edgefirst.studio/public/projects/2839/home) (80 classes). Part of the [EdgeFirst Model Zoo](https://huggingface.co/spaces/EdgeFirst/Models).
> [!TIP]
> **Training session**: [View on EdgeFirst Studio](https://test.edgefirst.studio/public/projects/2839/experiment/training/list?exp_id=4656) β dataset, training config, metrics, and exported artifacts.
> [!NOTE]
> Newer architecture with attention blocks.
---
## Size Comparison
All models validated on COCO val2017 (5000 images, 80 classes).
| Size | Params | GFLOPs | ONNX FP32 mAP@0.5 | ONNX FP32 mAP@0.5-0.95 | TFLite INT8 mAP@0.5 | TFLite INT8 mAP@0.5-0.95 |
|------|--------|--------|--------------------|-------------------------|----------------------|--------------------------|
| Nano | 2.6M | 6.5 | 53.4% | 37.9% | 50.1% | 34.5% |
| Small | 9.4M | 21.5 | β | β | β | β |
| Medium | 20.1M | 68.0 | β | β | β | β |
| Large | 25.3M | 87.6 | β | β | β | β |
| XLarge | 56.9M | 195.0 | β | β | β | β |
---
## On-Target Performance
Full pipeline timing: pre-processing + inference + post-processing.
| Size | Platform | Pre-proc (ms) | Inference (ms) | Post-proc (ms) | Total (ms) | FPS |
|------|----------|---------------|----------------|-----------------|------------|-----|
| β | β | β | β | β | β | β |
*Measured with [EdgeFirst Perception](https://github.com/EdgeFirstAI) stack. Timing includes full GStreamer pipeline overhead.*
---
## Downloads
<details open>
<summary><strong>ONNX FP32</strong> β Any platform with ONNX Runtime.</summary>
| Size | File | Status |
|------|------|--------|
| Nano | `yolo11n-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11n-det-coco.onnx) |
| Small | `yolo11s-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11s-det-coco.onnx) |
| Medium | `yolo11m-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11m-det-coco.onnx) |
| Large | `yolo11l-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11l-det-coco.onnx) |
| XLarge | `yolo11x-det-coco.onnx` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/onnx/yolo11x-det-coco.onnx) |
</details>
<details>
<summary><strong>TFLite INT8</strong> β CPU or NPU via runtime delegate (i.MX 8M Plus VX Delegate).</summary>
| Size | File | Status |
|------|------|--------|
| Nano | `yolo11n-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11n-det-coco.tflite) |
| Small | `yolo11s-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11s-det-coco.tflite) |
| Medium | `yolo11m-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11m-det-coco.tflite) |
| Large | `yolo11l-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11l-det-coco.tflite) |
| XLarge | `yolo11x-det-coco.tflite` | [Download](https://huggingface.co/EdgeFirst/yolo11-det/resolve/main/tflite/yolo11x-det-coco.tflite) |
</details>
---
## Deploy with EdgeFirst Perception
Copy-paste [GStreamer](https://github.com/EdgeFirstAI/gstreamer) pipeline examples for each platform.
### NXP i.MX 8M Plus β Camera to Detection with Vivante NPU
```bash
gst-launch-1.0 \
v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
edgefirstcameraadaptor ! \
tensor_filter framework=tensorflow-lite \
model=yolo11n-det-coco.tflite \
custom=Delegate:External,ExtDelegateLib:libvx_delegate.so ! \
edgefirstdetdecoder ! edgefirstoverlay ! waylandsink
```
### RPi5 + Hailo-8L
```bash
gst-launch-1.0 \
v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
hailonet hef-path=yolo11n-det-coco.hailo8l.hef ! \
hailofilter function-name=yolo11_nms ! \
hailooverlay ! videoconvert ! autovideosink
```
### NVIDIA Jetson (TensorRT)
```bash
gst-launch-1.0 \
v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
edgefirstcameraadaptor ! \
nvinfer config-file-path=yolo11n-det-coco-config.txt ! \
edgefirstdetdecoder ! edgefirstoverlay ! nveglglessink
```
*Full pipeline documentation: [EdgeFirst GStreamer Plugins](https://github.com/EdgeFirstAI/gstreamer)*
---
## Foundation (HAL) Python Integration
```python
from edgefirst.hal import Model, TensorImage
# Load model β metadata (labels, decoder config) is embedded in the file
model = Model("yolo11n-det-coco.tflite")
# Run inference on an image
image = TensorImage.from_file("image.jpg")
results = model.predict(image)
# Access detections
for det in results.detections:
print(f"{det.label}: {det.confidence:.2f} at {det.bbox}")
```
*[EdgeFirst HAL](https://github.com/EdgeFirstAI/hal) β Hardware abstraction layer with accelerated inference delegates.*
---
## CameraAdaptor
EdgeFirst [CameraAdaptor](https://github.com/EdgeFirstAI/cameraadaptor) enables training and inference directly on native sensor formats (GREY, YUYV, etc.) β skipping the ISP color conversion pipeline entirely. This reduces latency and power consumption on edge devices.
CameraAdaptor variants are included alongside baseline RGB models:
| Variant | Input Format | Use Case |
|---------|-------------|----------|
| `yolo11n-det-coco.onnx` | RGB (3ch) | Standard camera input |
| `yolo11n-det-coco-grey.onnx` | GREY (1ch) | Monochrome / IR sensors |
| `yolo11n-det-coco-yuyv.onnx` | YUYV (2ch) | Raw sensor bypass |
*Train CameraAdaptor models with [EdgeFirst Studio](https://edgefirst.studio) β the CameraAdaptor layer is automatically inserted during training.*
---
## Train Your Own with EdgeFirst Studio
Train on your own dataset with [**EdgeFirst Studio**](https://edgefirst.studio):
- **Free tier** includes YOLO training with automatic INT8 quantization and edge deployment
- Upload datasets via [EdgeFirst Recorder](https://github.com/EdgeFirstAI/recorder) or COCO/YOLO format
- AI-assisted annotation with auto-labeling
- CameraAdaptor integration for native sensor format training
- Deploy trained models to edge devices via [EdgeFirst Client](https://github.com/EdgeFirstAI/client)
---
## See Also
Other models in the [EdgeFirst Model Zoo](https://huggingface.co/spaces/EdgeFirst/Models):
| Model | Task | Best Nano Metric | Link |
|-------|------|-------------------|------|
| YOLOv5 Detection | Detection | 49.6% mAP@0.5 (ONNX) | [EdgeFirst/yolov5-det](https://huggingface.co/EdgeFirst/yolov5-det) |
| YOLOv8 Detection | Detection | 50.2% mAP@0.5 (ONNX) | [EdgeFirst/yolov8-det](https://huggingface.co/EdgeFirst/yolov8-det) |
| YOLOv8 Segmentation | Segmentation | 34.1% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolov8-seg](https://huggingface.co/EdgeFirst/yolov8-seg) |
| YOLO11 Segmentation | Segmentation | 35.5% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolo11-seg](https://huggingface.co/EdgeFirst/yolo11-seg) |
| YOLO26 Detection | Detection | 54.9% mAP@0.5 (ONNX) | [EdgeFirst/yolo26-det](https://huggingface.co/EdgeFirst/yolo26-det) |
| YOLO26 Segmentation | Segmentation | 37.0% Mask mAP@0.5-0.95 (ONNX) | [EdgeFirst/yolo26-seg](https://huggingface.co/EdgeFirst/yolo26-seg) |
---
## Technical Details
### Quantization Pipeline
All TFLite INT8 models are produced by EdgeFirst's custom quantization pipeline ([details](https://github.com/EdgeFirstAI/studio-ultralytics)):
1. **ONNX Export** β Standard Ultralytics export with `simplify=True`
2. **TF-Wrapped ONNX** β Box coordinates normalized to [0,1] inside DFL decode via `tf_wrapper` (~1.2% better mAP than post-hoc normalization)
3. **Split Decoder** β Boxes, scores, and mask coefficients split into separate output tensors for independent INT8 quantization scales
4. **Smart Calibration** β 500 images selected via greedy coverage maximization from COCO val2017
5. **Full INT8** β `uint8` input (raw pixels), `int8` output (per-tensor scales), MLIR quantizer
### Split Decoder Output Format
**Detection** (e.g., yolo11n):
- Boxes: `(1, 4, 8400)` β normalized [0,1] coordinates
- Scores: `(1, 80, 8400)` β class probabilities
Each tensor has independent quantization scale and zero-point. EdgeFirst HAL handles dequantization and reassembly automatically.
### Metadata
- **TFLite**: `edgefirst.json`, `labels.txt`, and `edgefirst.yaml` embedded via ZIP (no `tflite-support` dependency)
- **ONNX**: `edgefirst.json` embedded via `model.metadata_props`
No standalone metadata files β models are self-contained.
---
## Limitations
- **COCO bias** β Models trained on COCO (80 classes) inherit its biases: Western-centric scenes, specific object distributions, limited weather/lighting diversity
- **INT8 accuracy loss** β Full-integer quantization typically degrades mAP by 6-12% relative to FP32; actual loss depends on model architecture and dataset
- **Thermal variation** β On-target performance varies with device temperature; sustained inference may throttle on passively-cooled devices
- **Input resolution** β All models expect 640Γ640 input; other resolutions require letterboxing or may reduce accuracy
- **CameraAdaptor variants** β GREY/YUYV models trade color information for latency; accuracy may differ from RGB baseline depending on the task
---
## Citation
```bibtex
@software{edgefirst_yolo11_det,
title = { {YOLO11 Detection β EdgeFirst Edge AI} },
author = {Au-Zone Technologies},
url = {https://huggingface.co/EdgeFirst/yolo11-det},
year = {2026},
license = {Apache-2.0},
}
```
---
<p align="center">
<sub>
<a href="https://edgefirst.studio">EdgeFirst Studio</a> Β· <a href="https://github.com/EdgeFirstAI">GitHub</a> Β· <a href="https://doc.edgefirst.ai">Docs</a> Β· <a href="https://www.au-zone.com">Au-Zone Technologies</a><br>
Apache 2.0 Β· Β© Au-Zone Technologies Inc.
</sub>
</p>
|