sebastientaylor commited on
Commit
65914df
·
verified ·
1 Parent(s): 77e9014

EdgeFirst Model Zoo landing page

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ 01-ecosystem.png filter=lfs diff=lfs merge=lfs -text
37
+ 02-model-lifecycle.png filter=lfs diff=lfs merge=lfs -text
38
+ 03-on-target-validation.png filter=lfs diff=lfs merge=lfs -text
39
+ 04-coverage-matrix.png filter=lfs diff=lfs merge=lfs -text
01-ecosystem.png ADDED

Git LFS Details

  • SHA256: 3cd34c1a01de82e7834583c230766f526fc5829817402c82662a88bfded494a2
  • Pointer size: 131 Bytes
  • Size of remote file: 133 kB
01-ecosystem.svg ADDED
02-model-lifecycle.png ADDED

Git LFS Details

  • SHA256: a57d676b32277612f46698895de99a2c98d61143ab56a205b63164ed5076bb81
  • Pointer size: 131 Bytes
  • Size of remote file: 126 kB
03-on-target-validation.png ADDED

Git LFS Details

  • SHA256: dae552d4d0e9ec7c55ad4da2027d8e38b73aa1f8f5bc890484d38c8134ac219d
  • Pointer size: 131 Bytes
  • Size of remote file: 120 kB
04-coverage-matrix.png ADDED

Git LFS Details

  • SHA256: 26d22ff8b81dcf20ee1dc80ebc998ed947da0b4dfd8f62fbb5431bcab615eab8
  • Pointer size: 131 Bytes
  • Size of remote file: 123 kB
README.md CHANGED
@@ -1,10 +1,112 @@
1
  ---
2
- title: Models
3
- emoji: 🏢
4
  colorFrom: indigo
5
- colorTo: blue
6
  sdk: static
7
- pinned: false
 
8
  ---
9
 
10
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: EdgeFirst AI
3
+ emoji: 🔬
4
  colorFrom: indigo
5
+ colorTo: red
6
  sdk: static
7
+ pinned: true
8
+ license: apache-2.0
9
  ---
10
 
11
+ # EdgeFirst AI Spatial Perception at the Edge
12
+
13
+ **EdgeFirst Perception** is an open-source suite of libraries and microservices for AI-driven spatial perception on edge devices. It supports cameras, LiDAR, radar, and time-of-flight sensors — enabling real-time object detection, segmentation, sensor fusion, and 3D spatial understanding, optimized for resource-constrained embedded hardware.
14
+
15
+ [![EdgeFirst Studio](https://img.shields.io/badge/EdgeFirst_Studio-3E3371?style=for-the-badge&logoColor=white)](https://edgefirst.studio)
16
+ [![GitHub](https://img.shields.io/badge/GitHub-212529?style=for-the-badge&logo=github&logoColor=white)](https://github.com/EdgeFirstAI)
17
+ [![Documentation](https://img.shields.io/badge/Documentation-1FA0A8?style=for-the-badge&logo=readthedocs&logoColor=white)](https://doc.edgefirst.ai)
18
+ [![Au-Zone Technologies](https://img.shields.io/badge/Au--Zone_Technologies-6C757D?style=for-the-badge)](https://www.au-zone.com)
19
+
20
+ ---
21
+
22
+ ## Workflow
23
+
24
+ <img src="https://huggingface.co/spaces/EdgeFirst/README/resolve/main/01-ecosystem.png" alt="EdgeFirst Model Zoo Ecosystem"/>
25
+
26
+ Every model in the EdgeFirst Model Zoo passes through a validated pipeline. [**EdgeFirst Studio**](https://edgefirst.studio) manages datasets, training, multi-format export (ONNX, TFLite INT8, eIQ Neutron, Kinara DVM, HailoRT HEF, TensorRT), and reference validation. Models are then deployed to our board farm for **full-dataset on-target validation** on real hardware — measuring both accuracy (mAP) and detailed timing breakdown per device. Results are published here on HuggingFace with per-platform performance tables.
27
+
28
+ ## Model Lifecycle
29
+
30
+ <img src="https://huggingface.co/spaces/EdgeFirst/README/resolve/main/02-model-lifecycle.png" alt="Model Lifecycle: Training to Publication"/>
31
+
32
+ ## On-Target Validation
33
+
34
+ <img src="https://huggingface.co/spaces/EdgeFirst/README/resolve/main/03-on-target-validation.png" alt="On-Target Validation Pipeline"/>
35
+
36
+ Unlike desktop-only benchmarks, EdgeFirst validates every model on **real target hardware** with the full dataset. Each device produces both accuracy metrics (mAP) and a detailed timing breakdown — load, preprocessing, NPU inference, and decode — so you know exactly how a model performs on your specific platform.
37
+
38
+ ---
39
+
40
+ ## Supported Hardware
41
+
42
+ ![NXP i.MX 8M Plus](https://img.shields.io/badge/NXP-i.MX_8M_Plus-3E3371?style=flat-square&logoColor=white)
43
+ ![NXP i.MX 95](https://img.shields.io/badge/NXP-i.MX_95-3E3371?style=flat-square&logoColor=white)
44
+ ![NXP Ara240](https://img.shields.io/badge/NXP-Ara240-3E3371?style=flat-square&logoColor=white)
45
+ ![RPi5 + Hailo-8/8L](https://img.shields.io/badge/RPi5-Hailo--8%2F8L-1FA0A8?style=flat-square&logoColor=white)
46
+ ![NVIDIA Jetson](https://img.shields.io/badge/NVIDIA-Jetson-76B900?style=flat-square&logoColor=white)
47
+
48
+ ---
49
+
50
+ ## Model Zoo
51
+
52
+ Pre-trained YOLO models for edge deployment. Each model repo contains all sizes (nano through x-large), ONNX FP32 and TFLite INT8 formats, with platform-specific compiled variants as they become available.
53
+
54
+ ### Detection
55
+
56
+ | Model | Sizes | Nano mAP@0.5 | Link |
57
+ |-------|-------|-------------|------|
58
+ | **YOLO26** | n/s/m/l/x | 54.9% | [EdgeFirst/yolo26-det](https://huggingface.co/EdgeFirst/yolo26-det) |
59
+ | **YOLO11** | n/s/m/l/x | 53.4% | [EdgeFirst/yolo11-det](https://huggingface.co/EdgeFirst/yolo11-det) |
60
+ | **YOLOv8** | n/s/m/l/x | 50.2% | [EdgeFirst/yolov8-det](https://huggingface.co/EdgeFirst/yolov8-det) |
61
+ | **YOLOv5** | n/s/m/l/x | 49.6% | [EdgeFirst/yolov5-det](https://huggingface.co/EdgeFirst/yolov5-det) |
62
+
63
+ ### Instance Segmentation
64
+
65
+ | Model | Sizes | Nano Mask mAP | Link |
66
+ |-------|-------|--------------|------|
67
+ | **YOLO26** | n/s/m/l/x | 37.0% | [EdgeFirst/yolo26-seg](https://huggingface.co/EdgeFirst/yolo26-seg) |
68
+ | **YOLO11** | n/s/m/l/x | 35.5% | [EdgeFirst/yolo11-seg](https://huggingface.co/EdgeFirst/yolo11-seg) |
69
+ | **YOLOv8** | n/s/m/l/x | 34.1% | [EdgeFirst/yolov8-seg](https://huggingface.co/EdgeFirst/yolov8-seg) |
70
+
71
+ ---
72
+
73
+ ## Naming Convention
74
+
75
+ | Component | Pattern | Example |
76
+ |-----------|---------|---------|
77
+ | HF Repo | `EdgeFirst/{version}-{task}` | `EdgeFirst/yolov8-det` |
78
+ | ONNX Model | `{version}{size}-{task}-coco.onnx` | `yolov8n-det-coco.onnx` |
79
+ | TFLite Model | `{version}{size}-{task}-coco.tflite` | `yolov8n-det-coco.tflite` |
80
+ | i.MX 95 Model | `{version}{size}-{task}-coco.imx95.tflite` | `yolov8n-det-coco.imx95.tflite` |
81
+ | Studio Project | `{Dataset} {Task}` | `COCO Detection` |
82
+ | Studio Experiment | `{Version} {Task}` | `YOLOv8 Detection` |
83
+
84
+ ## Validation Pipeline
85
+
86
+ | Stage | What | Where |
87
+ |-------|------|-------|
88
+ | **Reference** | ONNX FP32 and TFLite INT8 mAP on full COCO val2017 (5000 images) | EdgeFirst Studio (cloud) |
89
+ | **On-Target** | Full dataset mAP + timing breakdown per device | Board farm (real hardware) |
90
+
91
+ ## Perception Architecture
92
+
93
+ | Layer | Description |
94
+ |-------|-------------|
95
+ | **Foundation** | Hardware abstraction, video I/O, accelerated inference delegates |
96
+ | **Zenoh** | Modular perception pipeline over Zenoh pub/sub |
97
+ | **GStreamer** | Spatial perception elements for GStreamer / NNStreamer |
98
+ | **ROS 2** | Native ROS 2 nodes extending Zenoh microservices *(Roadmap)* |
99
+
100
+ ## EdgeFirst Studio
101
+
102
+ [**EdgeFirst Studio**](https://edgefirst.studio) is the MLOps platform that drives the entire model zoo pipeline. **Free tier available.**
103
+
104
+ - Dataset management & AI-assisted annotation
105
+ - Model training with automatic multi-format export and INT8 quantization
106
+ - Reference and on-target validation with full metrics collection
107
+ - CameraAdaptor integration for native sensor format training
108
+ - Deploy trained models to edge devices via the [EdgeFirst Client](https://github.com/EdgeFirstAI/client) CLI
109
+
110
+ ---
111
+
112
+ Apache 2.0 · [Au-Zone Technologies Inc.](https://www.au-zone.com)
index.html CHANGED
@@ -1,19 +1,337 @@
1
- <!doctype html>
2
- <html>
3
- <head>
4
- <meta charset="utf-8" />
5
- <meta name="viewport" content="width=device-width" />
6
- <title>My static Space</title>
7
- <link rel="stylesheet" href="style.css" />
8
- </head>
9
- <body>
10
- <div class="card">
11
- <h1>Welcome to your static Space!</h1>
12
- <p>You can modify this app directly by editing <i>index.html</i> in the Files and versions tab.</p>
13
- <p>
14
- Also don't forget to check the
15
- <a href="https://huggingface.co/docs/hub/spaces" target="_blank">Spaces documentation</a>.
16
- </p>
17
- </div>
18
- </body>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  </html>
 
1
+ <!DOCTYPE html>
2
+ <html lang="en">
3
+ <head>
4
+ <meta charset="UTF-8">
5
+ <meta name="viewport" content="width=device-width, initial-scale=1.0">
6
+ <title>EdgeFirst AI — Spatial Perception at the Edge</title>
7
+ <base target="_blank">
8
+ <link rel="preconnect" href="https://fonts.googleapis.com">
9
+ <link rel="preconnect" href="https://fonts.gstatic.com" crossorigin>
10
+ <link href="https://fonts.googleapis.com/css2?family=Barlow:wght@300;400;500;600;700&family=Crimson+Text:wght@400;600&family=JetBrains+Mono:wght@400;500&display=swap" rel="stylesheet">
11
+ <style>
12
+ :root {
13
+ --navy: #3E3371;
14
+ --gold: #E8B820;
15
+ --teal: #1FA0A8;
16
+ --teal-text: #167A80;
17
+ --indigo: #4B0082;
18
+ --blue: #8FA3D4;
19
+ --bg: #FFFFFF;
20
+ --bg-subtle: #F8F9FA;
21
+ --bg-card: #F0EDF8;
22
+ --text: #343A40;
23
+ --text-strong: #212529;
24
+ --text-muted: #6C757D;
25
+ --border: #E9ECEF;
26
+ --heading: var(--navy);
27
+ --link: var(--teal-text);
28
+ }
29
+ @media (prefers-color-scheme: dark) {
30
+ :root {
31
+ --bg: #1a1a2e;
32
+ --bg-subtle: #16213e;
33
+ --bg-card: rgba(75, 0, 130, 0.2);
34
+ --text: #F1F3F5;
35
+ --text-strong: #FFFFFF;
36
+ --text-muted: #aaa;
37
+ --border: rgba(255,255,255,0.1);
38
+ --heading: #FFFFFF;
39
+ --link: #1FA0A8;
40
+ }
41
+ }
42
+ * { margin: 0; padding: 0; box-sizing: border-box; }
43
+ body {
44
+ font-family: 'Crimson Text', Georgia, serif;
45
+ background: var(--bg);
46
+ color: var(--text);
47
+ line-height: 1.7;
48
+ font-size: 16px;
49
+ }
50
+ .container { max-width: 860px; margin: 0 auto; padding: 2.5rem 2rem; }
51
+ h1, h2, h3 {
52
+ font-family: 'Barlow', -apple-system, sans-serif;
53
+ color: var(--heading);
54
+ line-height: 1.3;
55
+ }
56
+ h1 { font-size: 2.2rem; font-weight: 700; margin-bottom: 0.15rem; }
57
+ h1 .edge { color: var(--heading); }
58
+ h1 .first { color: var(--gold); }
59
+ h2 {
60
+ font-size: 1.15rem; font-weight: 600;
61
+ text-transform: uppercase; letter-spacing: 0.06em;
62
+ margin-top: 2.5rem; margin-bottom: 0.75rem;
63
+ padding-bottom: 0.35rem; border-bottom: 2px solid var(--gold);
64
+ }
65
+ .tagline {
66
+ font-family: 'Barlow', sans-serif; font-weight: 500;
67
+ letter-spacing: 0.12em; text-transform: uppercase;
68
+ color: var(--text-muted); font-size: 0.85rem; margin-bottom: 1.5rem;
69
+ }
70
+ p { margin-bottom: 1rem; }
71
+ a { color: var(--link); text-decoration: none; }
72
+ a:hover { text-decoration: underline; color: var(--navy); }
73
+ @media (prefers-color-scheme: dark) { a:hover { color: var(--gold); } }
74
+ .badges { display: flex; flex-wrap: wrap; gap: 0.4rem; margin: 1.25rem 0; }
75
+ .badges img { height: 22px; }
76
+ .link-badges { display: flex; flex-wrap: wrap; gap: 0.4rem; margin: 1.5rem 0; }
77
+ .link-badges img { height: 28px; }
78
+ /* Workflow diagram */
79
+ .diagram-container {
80
+ margin: 1.5rem 0;
81
+ text-align: center;
82
+ }
83
+ .diagram-container img {
84
+ max-width: 100%;
85
+ height: auto;
86
+ }
87
+ /* Naming table */
88
+ .naming-table {
89
+ width: 100%;
90
+ border-collapse: collapse;
91
+ margin: 0.75rem 0;
92
+ font-family: 'Barlow', sans-serif;
93
+ font-size: 0.9rem;
94
+ }
95
+ .naming-table th {
96
+ text-align: left;
97
+ padding: 0.4rem 0.6rem;
98
+ font-weight: 600;
99
+ font-size: 0.8rem;
100
+ text-transform: uppercase;
101
+ letter-spacing: 0.04em;
102
+ color: var(--text-muted);
103
+ border-bottom: 2px solid var(--border);
104
+ }
105
+ .naming-table td {
106
+ padding: 0.35rem 0.6rem;
107
+ border-bottom: 1px solid var(--border);
108
+ font-family: 'JetBrains Mono', monospace;
109
+ font-size: 0.82rem;
110
+ }
111
+ .naming-table td:first-child {
112
+ font-family: 'Barlow', sans-serif;
113
+ font-weight: 500;
114
+ color: var(--text-strong);
115
+ }
116
+ /* Architecture table */
117
+ .arch-table {
118
+ width: 100%;
119
+ border-collapse: collapse;
120
+ margin: 0.75rem 0;
121
+ font-family: 'Barlow', sans-serif;
122
+ font-size: 0.95rem;
123
+ }
124
+ .arch-table th {
125
+ text-align: left;
126
+ padding: 0.5rem 0.75rem;
127
+ font-weight: 600; font-size: 0.8rem;
128
+ text-transform: uppercase; letter-spacing: 0.05em;
129
+ color: var(--text-muted);
130
+ border-bottom: 2px solid var(--border);
131
+ }
132
+ .arch-table td {
133
+ padding: 0.5rem 0.75rem;
134
+ border-bottom: 1px solid var(--border);
135
+ }
136
+ .arch-table td:first-child {
137
+ font-weight: 600;
138
+ color: var(--text-strong);
139
+ white-space: nowrap;
140
+ }
141
+ .roadmap-tag {
142
+ font-family: 'Barlow', sans-serif;
143
+ font-size: 0.7rem; font-weight: 600;
144
+ padding: 0.1rem 0.45rem; border-radius: 3px;
145
+ background: var(--indigo); color: #fff;
146
+ vertical-align: middle; margin-left: 0.3rem;
147
+ letter-spacing: 0.03em; text-transform: uppercase;
148
+ }
149
+ .wip-tag {
150
+ font-family: 'Barlow', sans-serif;
151
+ font-size: 0.7rem; font-weight: 600;
152
+ padding: 0.1rem 0.45rem; border-radius: 3px;
153
+ background: var(--gold); color: #333;
154
+ vertical-align: middle; margin-left: 0.3rem;
155
+ letter-spacing: 0.03em; text-transform: uppercase;
156
+ }
157
+ /* Model sub-headings */
158
+ .container > h3 {
159
+ font-size: 0.95rem; font-weight: 600;
160
+ color: var(--text-muted);
161
+ text-transform: uppercase; letter-spacing: 0.05em;
162
+ margin-top: 1.25rem; margin-bottom: 0.5rem;
163
+ }
164
+ .model-grid {
165
+ display: grid;
166
+ grid-template-columns: repeat(auto-fill, minmax(240px, 1fr));
167
+ gap: 0.75rem; margin: 1rem 0;
168
+ }
169
+ .model-card {
170
+ font-family: 'Barlow', sans-serif;
171
+ background: var(--bg-card);
172
+ padding: 0.85rem 1rem; border-radius: 5px;
173
+ border-left: 3px solid var(--gold);
174
+ transition: border-color 0.15s;
175
+ }
176
+ .model-card:hover { border-left-color: var(--teal); }
177
+ .model-card h3 { font-size: 0.95rem; font-weight: 600; margin-bottom: 0.2rem; }
178
+ .model-card h3 a { color: var(--text-strong); }
179
+ .model-card h3 a:hover { color: var(--link); text-decoration: none; }
180
+ .model-card .meta { color: var(--text-muted); font-size: 0.82rem; font-weight: 400; }
181
+ .studio-features { margin: 0.75rem 0 0 1.25rem; color: var(--text); font-size: 0.95rem; }
182
+ .studio-features li { margin-bottom: 0.25rem; }
183
+ .footer {
184
+ margin-top: 3rem; padding-top: 1.5rem;
185
+ border-top: 1px solid var(--border);
186
+ text-align: center;
187
+ font-family: 'Barlow', sans-serif;
188
+ font-size: 0.8rem; color: var(--text-muted);
189
+ }
190
+ .footer a { color: var(--text-muted); }
191
+ .footer a:hover { color: var(--link); }
192
+ </style>
193
+ </head>
194
+ <body>
195
+ <div class="container">
196
+ <h1><span class="edge">Edge</span><span class="first">First</span> AI</h1>
197
+ <p class="tagline">AI for Spatial Perception</p>
198
+
199
+ <p>
200
+ <strong>EdgeFirst Perception</strong> is an open-source suite of libraries and microservices for AI-driven spatial perception on edge devices. It supports cameras, LiDAR, radar, and time-of-flight sensors &mdash; enabling real-time object detection, segmentation, sensor fusion, and 3D spatial understanding, optimized for resource-constrained embedded hardware.
201
+ </p>
202
+
203
+ <div class="link-badges">
204
+ <a href="https://edgefirst.studio"><img src="https://img.shields.io/badge/EdgeFirst_Studio-3E3371?style=for-the-badge&logoColor=white" alt="EdgeFirst Studio"></a>
205
+ <a href="https://github.com/EdgeFirstAI"><img src="https://img.shields.io/badge/GitHub-212529?style=for-the-badge&logo=github&logoColor=white" alt="GitHub"></a>
206
+ <a href="https://doc.edgefirst.ai"><img src="https://img.shields.io/badge/Documentation-1FA0A8?style=for-the-badge&logo=readthedocs&logoColor=white" alt="Documentation"></a>
207
+ <a href="https://www.au-zone.com"><img src="https://img.shields.io/badge/Au--Zone_Technologies-6C757D?style=for-the-badge" alt="Au-Zone Technologies"></a>
208
+ </div>
209
+
210
+ <h2>Workflow</h2>
211
+
212
+ <div class="diagram-container">
213
+ <img src="01-ecosystem.png" alt="EdgeFirst Model Zoo Ecosystem: Training, Validation, and Publication Workflow">
214
+ </div>
215
+
216
+ <p>
217
+ Every model in the EdgeFirst Model Zoo passes through a validated pipeline. <a href="https://edgefirst.studio"><strong>EdgeFirst Studio</strong></a> manages datasets, training, multi-format export (ONNX, TFLite INT8, eIQ Neutron, Kinara DVM, HailoRT HEF, TensorRT), and reference validation. Models are then deployed to our board farm for <strong>full-dataset on-target validation</strong> on real hardware &mdash; measuring both accuracy (mAP) and detailed timing breakdown per device. Results are published here on HuggingFace with per-platform performance tables.
218
+ </p>
219
+
220
+ <h2>Model Lifecycle</h2>
221
+
222
+ <div class="diagram-container">
223
+ <img src="02-model-lifecycle.png" alt="Model Lifecycle: 5 stages from training to publication">
224
+ </div>
225
+
226
+ <h2>On-Target Validation</h2>
227
+
228
+ <div class="diagram-container">
229
+ <img src="03-on-target-validation.png" alt="On-Target Validation Pipeline: full dataset validation on real hardware">
230
+ </div>
231
+
232
+ <p>
233
+ Unlike desktop-only benchmarks, EdgeFirst validates every model on <strong>real target hardware</strong> with the full dataset. Each device produces both accuracy metrics (mAP) and a detailed timing breakdown &mdash; load, preprocessing, NPU inference, and decode &mdash; so you know exactly how a model performs on your specific platform.
234
+ </p>
235
+
236
+ <h2>Supported Hardware</h2>
237
+ <div class="badges">
238
+ <img src="https://img.shields.io/badge/NXP-i.MX_8M_Plus-3E3371?style=flat-square&logoColor=white" alt="NXP i.MX 8M Plus">
239
+ <img src="https://img.shields.io/badge/NXP-i.MX_95-3E3371?style=flat-square&logoColor=white" alt="NXP i.MX 95">
240
+ <img src="https://img.shields.io/badge/NXP-Ara240-3E3371?style=flat-square&logoColor=white" alt="NXP Ara240">
241
+ <img src="https://img.shields.io/badge/RPi5-Hailo--8%2F8L-1FA0A8?style=flat-square&logoColor=white" alt="RPi5 + Hailo-8/8L">
242
+ <img src="https://img.shields.io/badge/NVIDIA-Jetson-76B900?style=flat-square&logoColor=white" alt="NVIDIA Jetson">
243
+ </div>
244
+
245
+ <h2>Model Zoo</h2>
246
+ <p>Pre-trained YOLO models for edge deployment. Each model repo contains all sizes (nano through x-large), ONNX FP32 and TFLite INT8 formats, with platform-specific compiled variants as they become available.</p>
247
+
248
+ <h3>Detection</h3>
249
+ <div class="model-grid">
250
+ <div class="model-card">
251
+ <h3><a href="https://huggingface.co/EdgeFirst/yolo26-det">YOLO26</a></h3>
252
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano mAP@0.5: 54.9%</p>
253
+ </div>
254
+ <div class="model-card">
255
+ <h3><a href="https://huggingface.co/EdgeFirst/yolo11-det">YOLO11</a></h3>
256
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano mAP@0.5: 53.4%</p>
257
+ </div>
258
+ <div class="model-card">
259
+ <h3><a href="https://huggingface.co/EdgeFirst/yolov8-det">YOLOv8</a></h3>
260
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano mAP@0.5: 50.2%</p>
261
+ </div>
262
+ <div class="model-card">
263
+ <h3><a href="https://huggingface.co/EdgeFirst/yolov5-det">YOLOv5</a></h3>
264
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano mAP@0.5: 49.6%</p>
265
+ </div>
266
+ </div>
267
+
268
+ <h3>Instance Segmentation</h3>
269
+ <div class="model-grid">
270
+ <div class="model-card">
271
+ <h3><a href="https://huggingface.co/EdgeFirst/yolo26-seg">YOLO26</a></h3>
272
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano Mask mAP: 37.0%</p>
273
+ </div>
274
+ <div class="model-card">
275
+ <h3><a href="https://huggingface.co/EdgeFirst/yolo11-seg">YOLO11</a></h3>
276
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano Mask mAP: 35.5%</p>
277
+ </div>
278
+ <div class="model-card">
279
+ <h3><a href="https://huggingface.co/EdgeFirst/yolov8-seg">YOLOv8</a></h3>
280
+ <p class="meta">n/s/m/l/x &middot; COCO 80 classes &middot; Nano Mask mAP: 34.1%</p>
281
+ </div>
282
+ </div>
283
+
284
+ <h2>Naming Convention</h2>
285
+ <p>Each HuggingFace repo contains one model family for one task, with all size variants inside.</p>
286
+ <table class="naming-table">
287
+ <tr><th>Component</th><th>Pattern</th><th>Example</th></tr>
288
+ <tr><td>HF Repo</td><td>EdgeFirst/{version}-{task}</td><td>EdgeFirst/yolov8-det</td></tr>
289
+ <tr><td>ONNX Model</td><td>{version}{size}-{task}-coco.onnx</td><td>yolov8n-det-coco.onnx</td></tr>
290
+ <tr><td>TFLite Model</td><td>{version}{size}-{task}-coco-int8.tflite</td><td>yolov8n-det-coco-int8.tflite</td></tr>
291
+ <tr><td>Studio Project</td><td>{Dataset} {Task}</td><td>COCO Detection</td></tr>
292
+ <tr><td>Studio Experiment</td><td>{Version} {Task}</td><td>YOLOv8 Detection</td></tr>
293
+ </table>
294
+
295
+ <h2>Validation Pipeline</h2>
296
+ <p>Models go through two validation stages before publication:</p>
297
+ <table class="arch-table">
298
+ <tr><th>Stage</th><th>What</th><th>Where</th></tr>
299
+ <tr>
300
+ <td>Reference</td>
301
+ <td>ONNX FP32 and TFLite INT8 mAP on full COCO val2017 (5000 images)</td>
302
+ <td>EdgeFirst Studio (cloud)</td>
303
+ </tr>
304
+ <tr>
305
+ <td>On-Target</td>
306
+ <td>Full dataset mAP + timing breakdown (load, preproc, invoke, decode, e2e) per device</td>
307
+ <td>Board farm (real hardware) <span class="wip-tag">In Progress</span></td>
308
+ </tr>
309
+ </table>
310
+
311
+ <h2>Perception Architecture</h2>
312
+ <table class="arch-table">
313
+ <tr><th>Layer</th><th>Description</th></tr>
314
+ <tr><td>Foundation</td><td>Hardware abstraction, video I/O, accelerated inference delegates</td></tr>
315
+ <tr><td>Zenoh</td><td>Modular perception pipeline over Zenoh pub/sub</td></tr>
316
+ <tr><td>GStreamer</td><td>Spatial perception elements for GStreamer / NNStreamer</td></tr>
317
+ <tr><td>ROS 2</td><td>Native ROS 2 nodes extending Zenoh microservices <span class="roadmap-tag">Roadmap</span></td></tr>
318
+ </table>
319
+
320
+ <h2>EdgeFirst Studio</h2>
321
+ <p>
322
+ <a href="https://edgefirst.studio"><strong>EdgeFirst Studio</strong></a> is the MLOps platform that drives the entire model zoo pipeline. <strong>Free tier available.</strong>
323
+ </p>
324
+ <ul class="studio-features">
325
+ <li>Dataset management &amp; AI-assisted annotation</li>
326
+ <li>Model training with automatic multi-format export and INT8 quantization</li>
327
+ <li>Reference and on-target validation with full metrics collection</li>
328
+ <li>CameraAdaptor integration for native sensor format training</li>
329
+ <li>Deploy trained models to edge devices via the <a href="https://github.com/EdgeFirstAI/client">EdgeFirst Client</a> CLI</li>
330
+ </ul>
331
+
332
+ <div class="footer">
333
+ <p>Apache 2.0 &middot; &copy; <a href="https://www.au-zone.com">Au-Zone Technologies Inc.</a></p>
334
+ </div>
335
+ </div>
336
+ </body>
337
  </html>