File size: 6,112 Bytes
6525449
 
 
 
0273ff7
 
6525449
0273ff7
6525449
0273ff7
6525449
0273ff7
6525449
 
 
0273ff7
 
 
 
 
 
 
 
6525449
 
 
 
 
 
0273ff7
6525449
0273ff7
 
 
 
 
 
9f53f34
0273ff7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6525449
 
 
0273ff7
 
 
 
 
 
 
6525449
 
 
 
 
0273ff7
 
 
6525449
 
 
0273ff7
6525449
0273ff7
6525449
0273ff7
 
 
 
 
 
 
 
 
6525449
 
 
 
 
0273ff7
 
 
 
6525449
 
 
 
 
 
 
 
 
0273ff7
 
6525449
0273ff7
 
6525449
 
 
 
 
0273ff7
6525449
0273ff7
6525449
 
 
0273ff7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147


# MicroAGI00: MicroAGI Egocentric Dataset (2025)

> **License:** MicroAGI00 Open Use, No-Resale v1.0 (see `LICENSE`).
> **No resale:** You may not sell or paywall this dataset or derivative data. Trained models/outputs may be released under any terms.

## What this is

MicroAGI00 is a large-scale **egocentric RGB+D** dataset of **human household manipulation**, aligned with the task style of the Stanford BEHAVIOR benchmark: [https://behavior.stanford.edu/challenge/index.html](https://behavior.stanford.edu/challenge/index.html)

It’s designed to be “robotics-ready” at the *signal level*: synchronized streams, clean packaging, strong QC, and consistent structure—so you can spend time modeling, not cleaning data.

## Quick facts

* **Modalities:** synchronized RGB + 16-bit depth + IMU
* **Resolution & rate (RGB):** 1920×1080 @ 30 FPS (in MCAP)
* **Depth:** 16-bit, losslessly compressed inside MCAP
* **Scale:** ≈1,000,000 synchronized RGB frames and ≈1,000,000 depth frames (≈1M frame pairs)
* **Container:** `.mcap` (all signals)
* **Previews:** for a subset of sequences, `.mp4` previews (annotated overlays / visualized depth for quick review)

> Note: MP4 previews may be lower quality than MCAP due to compression and post-processing. Research use should read from MCAP.

## What’s included per sequence

* One large **MCAP** file containing:

  * RGB frames (1080p/30 fps)
  * 16-bit depth stream (lossless compression)
  * IMU data (as available)

* **MP4 preview videos** (subset of sequences):

  * RGB preview (for quick visual QA)
  * Visualized depth preview (for quick visual QA)

## Labels / annotations  https://www.youtube.com/watch?v=4-RVKBj2bcw

The base MicroAGI00 release is **primarily raw synchronized signals** (RGB-D-IMU) and **does not ship with full-coverage labels**.

If you’ve seen demo videos with overlays: those demonstrate **what MicroAGI can produce** as an add-on (see below), not what is universally present in the base dump.

## Data quality and QC philosophy

MicroAGI00 is built around *trustworthy signal integrity*:

* Tight **RGB↔Depth synchronization** checks
* Automated detection and scoring of:

  * frame drops / time discontinuities
  * motion blur / exposure failures
  * depth sanity (range/invalid ratios), compression integrity
  * IMU continuity where available
* Consistent trimming and packaging, with **sequence-level quality ratings** to support filtering (e.g., “clean only” training vs. “wild” robustness training)

## Diversity and covariate-shift robustness

MicroAGI data is captured across **Europe and Asia**, intentionally spanning environments that create real-world distribution shift:

* different homes, layouts, lighting regimes, materials
* different hands/skins, tool choices, cultural cooking/object priors
* varied camera motions and operator styles

This is meant to be **covariate-shift resilient** data for models that need to generalize.

## Optional derived signals (available on request)

If you want more than raw RGB-D-IMU, MicroAGI can deliver *derived outputs* on top of the same sequences (or on newly captured data), such as:

* **Ego-motion / trajectories** (VIO-style)
* **SLAM reconstructions** (maps, trajectories, keyframes)
* **Accurate body pose estimation**
* **State-of-the-art 3D hand landmarks** (true 3D, not just 2D reprojections)
* Additional QA layers and consistency checks tailored to your training setup

These are provided as a service deliverable (and can be scoped to subsets / key frames / full coverage), depending on your needs.

## Data access and structure

* Each top-level sample folder typically contains:

  * an MCAP “raw dump” folder
  * an MCAP “processed/curated” folder (when applicable)
  * an `mp4/` previews folder (when available)

All authoritative signals are inside the **MCAP**. Use MP4s for fast browsing only.

## Getting started

* Inspect an MCAP: `mcap info your_sequence.mcap`
* Extract messages: `mcap cat --topics <topic> your_sequence.mcap > out.bin`
* Python readers: `pip install mcap` (see the MCAP Python docs) or any MCAP-compatible tooling.

Typical topics include RGB, depth, IMU, and any additional channels you may have requested.

## Intended uses

* Policy and skill learning (robotics / VLA)
* Action detection and segmentation
* Hand/pose estimation and grasp analysis (raw or with add-ons)
* Depth-based reconstruction, SLAM, scene understanding
* World-model pre/post training
* Robustness testing under real distribution shift

## Data rights, consent, and licensing options

All capture is **legally consented**, with **data rights documentation** attached. Depending on the engagement, rights can be structured as:

* **non-exclusive** usage rights (typical dataset access), or
* **exclusive** rights for specific task scopes / environments / cohorts (custom programs)

## Services and custom data

MicroAGI provides on-demand:

* New data capture via our operator network (Europe + Asia)
* ML-enhanced derived signals (ego-motion, pose, hands, SLAM)
* Real-to-Sim pipelines and robotics-ready packaging
* Custom QC gates to match your training/eval stack

Typical lead times: under two weeks (up to four weeks for large jobs).

## How to order more

Email `data@micro-agi.com` with:

* Task description
* Desired hours or frame counts
* Target environment constraints (if any)
* Rights preference (exclusive / non-exclusive)
* Proposed price

We reply within one business day with lead time and final pricing.

Questions: `info@micro-agi.com`

## License

This dataset is released under the **MicroAGI00 Open Use, No-Resale License v1.0** (custom). See [`LICENSE`](./LICENSE). Redistribution must be free-of-charge under the same license.

Required credit: **"This work uses the MicroAGI00 dataset (MicroAGI, 2025)."**

## Attribution reminder

Public uses of the Dataset or Derivative Data must include the credit line above in a reasonable location for the medium (papers, repos, product docs, dataset pages, demo descriptions). Attribution is appreciated but not required for Trained Models or Outputs.