File size: 6,112 Bytes
6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 9f53f34 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 6525449 0273ff7 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 |
# MicroAGI00: MicroAGI Egocentric Dataset (2025)
> **License:** MicroAGI00 Open Use, No-Resale v1.0 (see `LICENSE`).
> **No resale:** You may not sell or paywall this dataset or derivative data. Trained models/outputs may be released under any terms.
## What this is
MicroAGI00 is a large-scale **egocentric RGB+D** dataset of **human household manipulation**, aligned with the task style of the Stanford BEHAVIOR benchmark: [https://behavior.stanford.edu/challenge/index.html](https://behavior.stanford.edu/challenge/index.html)
It’s designed to be “robotics-ready” at the *signal level*: synchronized streams, clean packaging, strong QC, and consistent structure—so you can spend time modeling, not cleaning data.
## Quick facts
* **Modalities:** synchronized RGB + 16-bit depth + IMU
* **Resolution & rate (RGB):** 1920×1080 @ 30 FPS (in MCAP)
* **Depth:** 16-bit, losslessly compressed inside MCAP
* **Scale:** ≈1,000,000 synchronized RGB frames and ≈1,000,000 depth frames (≈1M frame pairs)
* **Container:** `.mcap` (all signals)
* **Previews:** for a subset of sequences, `.mp4` previews (annotated overlays / visualized depth for quick review)
> Note: MP4 previews may be lower quality than MCAP due to compression and post-processing. Research use should read from MCAP.
## What’s included per sequence
* One large **MCAP** file containing:
* RGB frames (1080p/30 fps)
* 16-bit depth stream (lossless compression)
* IMU data (as available)
* **MP4 preview videos** (subset of sequences):
* RGB preview (for quick visual QA)
* Visualized depth preview (for quick visual QA)
## Labels / annotations https://www.youtube.com/watch?v=4-RVKBj2bcw
The base MicroAGI00 release is **primarily raw synchronized signals** (RGB-D-IMU) and **does not ship with full-coverage labels**.
If you’ve seen demo videos with overlays: those demonstrate **what MicroAGI can produce** as an add-on (see below), not what is universally present in the base dump.
## Data quality and QC philosophy
MicroAGI00 is built around *trustworthy signal integrity*:
* Tight **RGB↔Depth synchronization** checks
* Automated detection and scoring of:
* frame drops / time discontinuities
* motion blur / exposure failures
* depth sanity (range/invalid ratios), compression integrity
* IMU continuity where available
* Consistent trimming and packaging, with **sequence-level quality ratings** to support filtering (e.g., “clean only” training vs. “wild” robustness training)
## Diversity and covariate-shift robustness
MicroAGI data is captured across **Europe and Asia**, intentionally spanning environments that create real-world distribution shift:
* different homes, layouts, lighting regimes, materials
* different hands/skins, tool choices, cultural cooking/object priors
* varied camera motions and operator styles
This is meant to be **covariate-shift resilient** data for models that need to generalize.
## Optional derived signals (available on request)
If you want more than raw RGB-D-IMU, MicroAGI can deliver *derived outputs* on top of the same sequences (or on newly captured data), such as:
* **Ego-motion / trajectories** (VIO-style)
* **SLAM reconstructions** (maps, trajectories, keyframes)
* **Accurate body pose estimation**
* **State-of-the-art 3D hand landmarks** (true 3D, not just 2D reprojections)
* Additional QA layers and consistency checks tailored to your training setup
These are provided as a service deliverable (and can be scoped to subsets / key frames / full coverage), depending on your needs.
## Data access and structure
* Each top-level sample folder typically contains:
* an MCAP “raw dump” folder
* an MCAP “processed/curated” folder (when applicable)
* an `mp4/` previews folder (when available)
All authoritative signals are inside the **MCAP**. Use MP4s for fast browsing only.
## Getting started
* Inspect an MCAP: `mcap info your_sequence.mcap`
* Extract messages: `mcap cat --topics <topic> your_sequence.mcap > out.bin`
* Python readers: `pip install mcap` (see the MCAP Python docs) or any MCAP-compatible tooling.
Typical topics include RGB, depth, IMU, and any additional channels you may have requested.
## Intended uses
* Policy and skill learning (robotics / VLA)
* Action detection and segmentation
* Hand/pose estimation and grasp analysis (raw or with add-ons)
* Depth-based reconstruction, SLAM, scene understanding
* World-model pre/post training
* Robustness testing under real distribution shift
## Data rights, consent, and licensing options
All capture is **legally consented**, with **data rights documentation** attached. Depending on the engagement, rights can be structured as:
* **non-exclusive** usage rights (typical dataset access), or
* **exclusive** rights for specific task scopes / environments / cohorts (custom programs)
## Services and custom data
MicroAGI provides on-demand:
* New data capture via our operator network (Europe + Asia)
* ML-enhanced derived signals (ego-motion, pose, hands, SLAM)
* Real-to-Sim pipelines and robotics-ready packaging
* Custom QC gates to match your training/eval stack
Typical lead times: under two weeks (up to four weeks for large jobs).
## How to order more
Email `data@micro-agi.com` with:
* Task description
* Desired hours or frame counts
* Target environment constraints (if any)
* Rights preference (exclusive / non-exclusive)
* Proposed price
We reply within one business day with lead time and final pricing.
Questions: `info@micro-agi.com`
## License
This dataset is released under the **MicroAGI00 Open Use, No-Resale License v1.0** (custom). See [`LICENSE`](./LICENSE). Redistribution must be free-of-charge under the same license.
Required credit: **"This work uses the MicroAGI00 dataset (MicroAGI, 2025)."**
## Attribution reminder
Public uses of the Dataset or Derivative Data must include the credit line above in a reasonable location for the medium (papers, repos, product docs, dataset pages, demo descriptions). Attribution is appreciated but not required for Trained Models or Outputs.
|