MCAP-Housing / README.md
shas232's picture
Update dataset card: remove device refs, add scaling and QC info
5da9b1c verified
---
license: cc-by-nc-4.0
task_categories:
- robotics
- video-classification
- depth-estimation
tags:
- egocentric
- rgbd
- manipulation
- mcap
- ros2
- imu
- pointcloud
- depth
- kitchen
- household
- slam
- hand-tracking
- body-pose
pretty_name: "MCAP-Housing: Egocentric RGB-D Manipulation Dataset"
size_categories:
- 1K<n<10K
---
# MCAP-Housing: Egocentric RGB-D Household Manipulation Dataset
**MCAP-Housing** is an egocentric **RGB + Depth + IMU** dataset of human household manipulation activities, packaged in robotics-native `.mcap` (ROS2) format. Designed for robotics research, policy learning, and embodied AI.
This is a **sample release**. We can scale to custom episode counts, new activities, and specific environments on request. Contact us to discuss your requirements.
---
## Quick Facts
| Property | Value |
|----------|-------|
| **Modalities** | Synchronized RGB + 16-bit Depth + IMU + Point Clouds |
| **Resolution (RGB)** | 1920 × 1440 @ 60 FPS |
| **Depth** | 16-bit millimeter, LiDAR-sourced, aligned to RGB |
| **Point Clouds** | Per-frame colored XYZRGB (up to 50k points) |
| **IMU** | 6-axis (accel + gyro) + magnetometer + gravity + orientation @ 60 Hz |
| **Pose** | 6DoF camera pose (world → camera transform) per frame |
| **Activities** | 10 household manipulation sequences |
| **Total Frames** | ~30,000 synchronized RGB-D pairs |
| **Total Size** | ~30 GB |
| **Container** | `.mcap` with ROS2 CDR serialization |
---
## What's Included Per Sequence
Each `.mcap` file contains **11 synchronized ROS2 topics**:
| Topic | Message Type | Description |
|-------|-------------|-------------|
| `/camera/rgb/compressed` | `sensor_msgs/CompressedImage` | JPEG-encoded RGB frames |
| `/camera/depth/aligned` | `sensor_msgs/Image` | Raw 16-bit depth aligned to RGB |
| `/camera/depth/filtered` | `sensor_msgs/Image` | Bilateral-filtered depth (hole-filled) |
| `/camera/depth/colorized` | `sensor_msgs/Image` | Turbo-colormap depth visualization |
| `/camera/points` | `sensor_msgs/PointCloud2` | Colored XYZRGB point cloud |
| `/camera/camera_info` | `sensor_msgs/CameraInfo` | Per-frame intrinsics (fx, fy, cx, cy) |
| `/tf` | `tf2_msgs/TFMessage` | 6DoF camera pose (world → camera) |
| `/imu` | `sensor_msgs/Imu` | Linear acceleration + angular velocity |
| `/imu/gravity` | `geometry_msgs/Vector3Stamped` | Gravity vector |
| `/imu/orientation` | `geometry_msgs/QuaternionStamped` | Device orientation quaternion |
| `/imu/mag` | `sensor_msgs/MagneticField` | Magnetometer readings |
---
## Available on Request
Beyond the raw synchronized streams, the following are available on request:
- **Ego-motion / trajectories** (VIO-style) — smooth, drift-corrected camera trajectories
- **SLAM reconstructions** — dense maps, optimized trajectories, keyframe selection
- **Accurate body pose estimation** — full skeletal tracking during manipulation
- **State-of-the-art 3D hand landmarks** — true 3D hand joint positions, not 2D reprojections
- **QC-validated data** — quality-checked sequences with automated scoring for frame drops, motion blur, depth sanity, and sync integrity
- **Additional QA layers and consistency checks** tailored to your specific training setup
Contact us to discuss which derived signals you need.
---
## Data Quality
- Tight RGB ↔ Depth ↔ IMU synchronization (all streams at 60 Hz)
- Per-frame camera intrinsics (not a single fixed calibration)
- Per-frame 6DoF pose from visual-inertial odometry
- Depth hole-filling and bilateral filtering provided as separate topics
- Full QC reports and filtered datasets available on request
---
## Getting Started
### Inspect a file
```bash
pip install mcap
mcap info Chopping.mcap
```
### Read in Python
```python
from mcap.reader import make_reader
from mcap_ros2.decoder import DecoderFactory
with open("Chopping.mcap", "rb") as f:
reader = make_reader(f, decoder_factories=[DecoderFactory()])
for schema, channel, message, decoded in reader.iter_decoded_messages():
if channel.topic == "/camera/rgb/compressed":
print(f"RGB frame at t={message.log_time}, size={len(decoded.data)} bytes")
elif channel.topic == "/camera/depth/aligned":
print(f"Depth frame: {decoded.width}x{decoded.height}, encoding={decoded.encoding}")
elif channel.topic == "/camera/points":
print(f"Point cloud: {decoded.width} points")
```
### Visualize
Open any `.mcap` file directly in [Foxglove Studio](https://foxglove.dev/) for full 3D visualization of RGB, depth, point clouds, and transforms.
### Dependencies
```bash
pip install mcap mcap-ros2-support numpy opencv-python
```
---
## Intended Uses
- **Policy and skill learning** — imitation learning, VLA pre-training
- **Action detection and segmentation** — temporal activity recognition
- **Hand and body pose estimation** — grasp analysis, manipulation understanding
- **Depth-based reconstruction** — SLAM, scene understanding, 3D mapping
- **World-model training** — ego-motion prediction, scene dynamics
- **Sensor fusion research** — RGB-D-IMU alignment and calibration
---
## Scaling & Custom Data
This release is a sample. We offer:
- **Custom episode capture** — specific activities, environments, and object sets
- **Scalable data collection** — hundreds to thousands of episodes on demand
- **Derived signal pipelines** — hand tracking, body pose, SLAM, tailored to your model
- **Custom QC gates** — filtering and validation matched to your training requirements
Reach out to discuss your needs.
---
## License
This dataset is released under **CC-BY-NC-4.0**. Free for research and non-commercial use with attribution. For commercial licensing, contact us.
**Required attribution:** *"This work uses the MCAP-Housing dataset (Cortex Data Labs, 2025)."*
---
## Contact
- **Email:** shashin.bhaskar@gmail.com
- **Organization:** [Cortex Data Labs](https://huggingface.co/cortexdatalabs)
For custom data capture, derived signals, QC-validated datasets, or commercial licensing, reach out directly.