| --- |
| license: apache-2.0 |
| task_categories: |
| - robotics |
| tags: |
| - LeRobot |
| - robotics |
| - egocentric |
| - hand-tracking |
| - manus-gloves |
| - imitation-learning |
| --- |
| |
| # manus-egocentric-sample |
|
|
| Egocentric video dataset with Manus glove hand tracking data, converted to LeRobot v3.0 format. |
|
|
| ## Dataset Description |
|
|
| This dataset contains egocentric (first-person view) recordings of human hands performing various manipulation tasks, captured with: |
| - **Manus Metagloves**: High-precision finger tracking (~70Hz) |
| - **OAK-D Camera**: RGB video (1920x1080, 30fps) + Depth (640x400, 30fps) |
| - **IMU**: Accelerometer and gyroscope data |
|
|
| ### Dataset Statistics |
|
|
| | Property | Value | |
| |----------|-------| |
| | Total Episodes | 20 | |
| | Total Frames | 96,544 | |
| | FPS | 30 | |
| | Robot Type | manus_gloves | |
| | LeRobot Version | v2.1 | |
| | Dataset Size | ~15.3 GB | |
| |
| ### Tasks |
| |
| `Pick_and_pack_task`, `fold_laundry` |
| |
| ### Features |
| |
| | Feature | Type | Shape | |
| |---------|------|-------| |
| | `episode_index` | int64 | [1] | |
| | `frame_index` | int64 | [1] | |
| | `timestamp` | float32 | [1] | |
| | `task_index` | int64 | [1] | |
| | `observation.images.egocentric` | video | [1080, 1920, 3] | |
| | `observation.state.hand_joints` | float32 | [150] | |
| | `observation.state.finger_angles` | float32 | [40] | |
| | `observation.state.gestures` | float32 | [26] | |
| | `observation.state.imu` | float32 | [6] | |
|
|
|
|
| ### Depth Data |
|
|
| This dataset includes raw depth data as a custom extension (LeRobot v3.0 doesn't officially support depth yet). |
|
|
| | Property | Value | |
| |----------|-------| |
| | Format | Raw uint16 binary | |
| | Resolution | [400, 640] (H, W) | |
| | Unit | millimeters | |
| | Episodes with depth | 17 | |
| | Storage | Episode-based (`depth/episode_XXXXXX.bin`) | |
|
|
| To load depth data: |
| ```python |
| import numpy as np |
| |
| def load_depth(dataset_path, episode_index): |
| depth_path = dataset_path / f"depth/episode_{episode_index:06d}.bin" |
| if depth_path.exists(): |
| data = np.fromfile(depth_path, dtype=np.uint16) |
| # Reshape based on frame count (need timestamps) |
| return data |
| return None |
| ``` |
|
|
|
|
| ## Usage |
|
|
| ### Load with LeRobot |
|
|
| ```python |
| from lerobot.datasets.lerobot_dataset import LeRobotDataset |
| |
| dataset = LeRobotDataset("opengraph-labs/manus-egocentric-sample") |
| |
| # Access a sample |
| sample = dataset[0] |
| print(sample.keys()) |
| # dict_keys(['observation.images.egocentric', 'observation.state.hand_joints', ...]) |
| |
| # Get hand joint positions (25 joints x 3 coords x 2 hands = 150 dims) |
| hand_joints = sample["observation.state.hand_joints"] |
| print(hand_joints.shape) # torch.Size([150]) |
| ``` |
|
|
| ### Data Structure |
|
|
| ``` |
| manus-egocentric-sample/ |
| ├── meta/ |
| │ ├── info.json # Dataset schema and features |
| │ ├── stats.json # Feature statistics (mean/std/min/max) |
| │ ├── tasks.jsonl # Task definitions |
| │ ├── episodes.jsonl # Episode metadata |
| │ └── episodes/ # Episode info (parquet) |
| ├── data/chunk-XXX/ # Time-series data (parquet) |
| ├── videos/ # RGB video (mp4) |
| └── depth/ # Raw depth data (uint16 binary) |
| ``` |
|
|
| ## Hand Skeleton Structure |
|
|
| Each hand has 25 tracked joints: |
|
|
| ``` |
| Hand (Root) |
| ├── Thumb: MCP → PIP → DIP → TIP (4 joints) |
| ├── Index: MCP → IP → PIP → DIP → TIP (5 joints) |
| ├── Middle: MCP → IP → PIP → DIP → TIP (5 joints) |
| ├── Ring: MCP → IP → PIP → DIP → TIP (5 joints) |
| └── Pinky: MCP → IP → PIP → DIP → TIP (5 joints) |
| ``` |
|
|
| ## Citation |
|
|
| If you use this dataset, please cite: |
|
|
| ```bibtex |
| @dataset{manus_egocentric_2025, |
| title = {Manus Egocentric Hand Tracking Dataset}, |
| author = {OpenGraph Labs}, |
| year = {2025}, |
| publisher = {Hugging Face}, |
| url = {https://huggingface.co/datasets/opengraph-labs/manus-egocentric-sample} |
| } |
| ``` |
|
|
| ## License |
|
|
| This dataset is released under the Apache 2.0 License. |
|
|