MicroAGI commited on
Commit
6525449
·
verified ·
1 Parent(s): 4509594

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +163 -5
README.md CHANGED
@@ -1,5 +1,163 @@
1
- ---
2
- license: other
3
- license_name: microagi-os-l1
4
- license_link: LICENSE
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: microagi-os-l1
4
+ license_link: LICENSE
5
+ task_categories:
6
+ - text-classification
7
+ language:
8
+ - en
9
+ tags:
10
+ - dataset
11
+ - egocentric
12
+ - robotics
13
+ - rgbd
14
+ - depth
15
+ - manipulation
16
+ - mcap
17
+ - ros2
18
+ - computer_vision
19
+ pretty_name: MicroAGI00 Egocentric Dataset for Simple Household Manipulation
20
+ size_categories:
21
+ - 1M<n<10M
22
+ ---
23
+ ---
24
+
25
+
26
+
27
+ # MicroAGI00: MicroAGI Egocentric Dataset (2025)
28
+
29
+ > Required credit: "This work uses the MicroAGI00 dataset (MicroAGI, 2025)."
30
+ > License: MicroAGI00 Open Use, No-Resale v1.0 (see `LICENSE`).
31
+ > No resale: You may not sell or paywall this dataset or derivative data. Trained models/outputs may be released under any terms.
32
+
33
+ ## Overview
34
+ MicroAGI00 is a large-scale egocentric RGB+D dataset of human manipulation in https://behavior.stanford.edu/challenge/index.html tasks.
35
+
36
+
37
+
38
+ ## Quick facts
39
+
40
+ * Modality: synchronized RGB + 16‑bit depth + IMU + annotations
41
+ * Resolution & rate (RGB): 1920×1080 @ 30 FPS (in MCAP)
42
+ * Depth: 16‑bit, losslessly compressed inside MCAP
43
+ * Scale: ≈1,000,000 synchronized RGB frames and ≈1,000,000 depth frames (≈1M frame pairs)
44
+ * Container: `.mcap` (all signals + annotations)
45
+ * Previews: For as sample for only some bags `.mp4` per sequence (annotated RGB; visualized native depth)
46
+ * Annotations: Only in %5 of the dataset, hand landmarks and short action text
47
+
48
+ ## What’s included per sequence
49
+
50
+ * One large **MCAP** file containing:
51
+
52
+ * RGB frames (1080p/30 fps)
53
+ * 16‑bit depth stream (lossless compression)
54
+ * IMU data (as available)
55
+ * For Some Data the Embedded annotations (hands, action text)
56
+
57
+ **MP4** preview videos:
58
+
59
+ * Annotated RGB (for quick review)
60
+ * Visualized native depth map (for quick review)
61
+
62
+ > Note: MP4 previews may be lower quality than MCAP due to compression and post‑processing. Research use should read from MCAP.
63
+
64
+ ## Annotations
65
+
66
+ Annotations are generated by our in‑house.
67
+
68
+ ### Hand annotations (per frame) — JSON schema example
69
+
70
+ ```
71
+ {
72
+ "frame_number": 9,
73
+ "timestamp_seconds": 0.3,
74
+ "resolution": { "width": 1920, "height": 1080 },
75
+ "hands": [
76
+ {
77
+ "hand_index": 0,
78
+ "landmarks": [
79
+ { "id": 0, "name": "WRIST", "x": 0.7124036550521851, "y": 0.7347621917724609, "z": -1.444301744868426e-07, "visibility": 0.0 },
80
+
81
+ ],
82
+ "hand": "Left",
83
+ "confidence": 0.9268525838851929
84
+ },
85
+ {
86
+ "hand_index": 1,
87
+ "landmarks": [
88
+ { "id": 0, "name": "WRIST", "x": 0.4461262822151184, "y": 0.35183972120285034, "z": -1.2342320587777067e-07, "visibility": 0.0 },
89
+
90
+ "hand": "Right",
91
+ "confidence": 0.908446729183197
92
+ }
93
+ ],
94
+ "frame_idx": 9,
95
+ "exact_frame_timestamp": 1758122341583104000,
96
+ "exact_frame_timestamp_sec": 1758122341.583104
97
+ }
98
+ ```
99
+
100
+ ### Text (action) annotations (per frame/window) — JSON schema example
101
+
102
+ ```
103
+ {
104
+ "schema_version": "v1.0",
105
+ "action_text": "Right hand, holding a knife, is chopping cooked meat held by the left hand on the red cutting board.",
106
+ "confidence": 1.0,
107
+ "source": { "model": "MicroAGI, MAGI01" },
108
+ "exact_frame_timestamp": 1758122341583104000,
109
+ "exact_frame_timestamp_sec": 1758122341.583104
110
+ }
111
+ ```
112
+
113
+ ## Data access and structure
114
+
115
+ * Each top-level sample folder contains: One folder of strong heavy mcap dump, one folder of annotated mcap dump, one folder of mp4 previews
116
+ * All authoritative signals and annotations are inside the MCAP. Use the MP4s for quick visual QA only.
117
+
118
+ ## Getting started
119
+
120
+ * Inspect an MCAP: `mcap info your_sequence.mcap`
121
+ * Extract messages: `mcap cat --topics <topic> your_sequence.mcap > out.bin`
122
+ * Python readers: `pip install mcap` (see the MCAP Python docs) or any MCAP-compatible tooling. Typical topics include RGB, depth, IMU, and annotation channels.
123
+
124
+ ## Intended uses
125
+
126
+ * Policy and skill learning (robotics/VLA)
127
+ * Action detection and segmentation
128
+ * Hand/pose estimation and grasp analysis
129
+ * Depth-based reconstruction, SLAM, scene understanding
130
+ * World-model pretraining
131
+
132
+ ## Services and custom data
133
+
134
+ MicroAGI provides on-demand:
135
+
136
+ * Real‑to‑Sim pipelines
137
+ * ML‑enhanced 3D point clouds and SLAM reconstructions
138
+ * New data capture via our network of skilled tradespeople and factory workers (often below typical market cost)
139
+ * Enablement for your workforce to wear our device and run through our processing pipeline
140
+
141
+ Typical lead times: under two weeks (up to four weeks for large jobs).
142
+
143
+ ## How to order more
144
+
145
+ Email `data@micro-agi.com` with:
146
+
147
+ * Task description
148
+ * Desired hours or frame counts
149
+ * Proposed price
150
+ We will reply within one business day with lead time and final pricing.
151
+
152
+ Questions: `info@micro-agi.com`
153
+
154
+ ## License
155
+
156
+ This dataset is released under the MicroAGI00 Open Use, No‑Resale License v1.0 (custom). See [`LICENSE`](./LICENSE). Redistribution must be free‑of‑charge under the same license. Required credit: "This work uses the MicroAGI00 dataset (MicroAGI, 2025)."
157
+
158
+
159
+
160
+
161
+ ## Attribution reminder
162
+
163
+ Public uses of the Dataset or Derivative Data must include the credit line above in a reasonable location for the medium (papers, repos, product docs, dataset pages, demo descriptions). Attribution is appreciated but not required for Trained Models or Outputs.