RealSource-World / README.md
Jacky0923's picture
Update README.md
75e3df7 verified
---
license: cc-by-nc-sa-4.0
pretty_name: RealSource World
size_categories:
- 10M<n<100M
task_categories:
- robotics
language:
- en
- zh
tags:
- robotics
- real-world
- dual-arm
- whole body control
- manipulation
extra_gated_prompt: >-
### REALSOURCE WORLD COMMUNITY LICENSE AGREEMENT: All the data and code
within this repo are under [CC BY-NC-SA
4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/).
extra_gated_fields:
First Name: text
Last Name: text
Email: text
Country: country
Affiliation: text
Phone: text
Job title:
type: select
options:
- Student
- Research Graduate
- AI researcher
- AI developer/engineer
- Reporter
- Other
geo: ip_location
By clicking Submit below I accept the terms of the license and acknowledge that the information I provide will be collected stored processed and shared in accordance with the RealSource Privacy Policy: checkbox
extra_gated_description: >-
The information you provide will be collected, stored, processed and shared in
accordance with the RealSource Privacy Policy.
extra_gated_button_content: Submit
---
<div align="center">
<video controls autoplay src="https://realmanrobot.github.io/real_source_dataset/assets/real_source_video-CQfv30ls.mp4"></video>
</div>
# RealSource World
RealSource World is a large-scale real-world robotics manipulation dataset collected using RS-02 dual-arm humanoid robot. This dataset contains diverse long-horizon manipulation tasks performed in real-world environments, with detailed annotations for atomic skills and quality assessments.
## Key Features
- **14+ million** frames of real-world dual-arm manipulation demonstrations.
- **11,428+** episodes across **35** distinct manipulation tasks.
- **71-dimensional** proprioceptive state space including joint positions, velocities, accelerations, forces, torques, and end-effector poses.
- **Multi-camera** visual observations (head camera, left hand camera, right hand camera) at 720Γ—1280 resolution, 30 FPS.
- **Fine-grained annotations** with atomic skill segmentation and quality assessments for each episode.
- **Diverse scenes** including kitchen, conference room, convenience store, and household environments.
- **Dual-arm coordination** tasks demonstrating complex bimanual manipulation skills.
## News
- **`[2025/12]`** RealSource World dataset fully uploaded with 35 tasks. [Download Link](https://huggingface.co/datasets/RealSourceData/RealSource-World)
- **`[2025/11]`** RealSource World released on Hugging Face. [Download Link](https://huggingface.co/datasets/RealSourceData/RealSource-world)
## Changelog
### Version 1.2 (December 2025)
- **Data Format Upgrade & Fixes**
- Upgraded codebase_version to v2.1
- Extended proprioceptive state space from 57 to **71 dimensions**
- **Added joint acceleration data (LeftJoint_Acc/RightJoint_Acc): 14 dimensions (7 Γ— 2)**
- Fixed timing synchronization issues in some data
- Optimized data structure for improved completeness and consistency
- All datasets reprocessed for high-quality output
### Version 1.1 (December 2025)
- **Full Dataset Upload**
- Complete release of all dataset files on Hugging Face
- 35 manipulation tasks with 11,428 episodes
- Total size: ~525GB
- Total files: ~104,000+
### Version 1.0 (November 2025)
- **Initial Release**
- Released RealSource World dataset on Hugging Face
- 29 manipulation tasks with 11,428 episodes
- 14+ million frames of real-world dual-arm manipulation demonstrations
- 57-dimensional proprioceptive state space
- Multi-camera visual observations (head, left hand, right hand)
- Fine-grained annotations with atomic skill segmentation
- Complete camera parameters (intrinsics and extrinsics) for all episodes
- Quality assessment for each episode
## Table of Contents
- [Key Features](#key-features)
- [News](#news)
- [Changelog](#changelog)
- [Quick Start](#quick-start)
- [Download Dataset](#download-dataset)
- [Dataset Structure](#dataset-structure)
- [Understanding the Dataset Format](#understanding-the-dataset-format)
- [Loading and Using the Dataset](#loading-and-using-the-dataset)
- [Data Format Details](#data-format-details)
- [Proprioceptive State (71 Dimensions)](#proprioceptive-state-71-dimensions)
- [Action Space (17 Dimensions)](#action-space-17-dimensions)
- [Visual Observations](#visual-observations)
- [Camera Parameters](#camera-parameters)
- [Sub-task Annotations](#sub-task-annotations)
- [Dataset Statistics](#dataset-statistics)
- [Robot URDF Model](#robot-urdf-model)
- [License and Citation](#license-and-citation)
## Quick Start
### Download Dataset
To download the complete dataset, you can use the following commands. If you encounter any issues, please refer to the Hugging Face official documentation.
```bash
# Ensure git-lfs is installed (https://git-lfs.com)
git lfs install
# When prompted for password, use an access token with read permissions.
# Generate from your settings page: https://huggingface.co/settings/tokens
git clone https://huggingface.co/datasets/RealSourceData/RealSource-World
# If you want to download only file pointers without large files
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/RealSourceData/RealSource-World
```
If you only want to download a specific task from the RealSource World dataset, for example `Arrange_the_cups`, follow these steps:
```bash
# Ensure Git LFS is installed (https://git-lfs.com)
git lfs install
# Initialize an empty Git repository
git init RealSource-World
cd RealSource-World
# Set remote repository
git remote add origin https://huggingface.co/datasets/RealSourceData/RealSource-World
# Enable sparse checkout
git sparse-checkout init
# Specify folders and files to download
git sparse-checkout set Arrange_the_cups scripts
# Pull data from main branch
git pull origin main
```
## Dataset Structure
### Folder Hierarchy
```
RealSource-World/
β”œβ”€β”€ Arrange_the_cups/
β”‚ β”œβ”€β”€ data/
β”‚ β”‚ └── chunk-000/
β”‚ β”‚ β”œβ”€β”€ episode_000000.parquet
β”‚ β”‚ β”œβ”€β”€ episode_000001.parquet
β”‚ β”‚ └── ...
β”‚ β”œβ”€β”€ meta/
β”‚ β”‚ β”œβ”€β”€ info.json # Dataset metadata and feature definitions
β”‚ β”‚ β”œβ”€β”€ episodes.jsonl # Episode-level metadata
β”‚ β”‚ β”œβ”€β”€ episodes_stats.jsonl # Episode statistics
β”‚ β”‚ β”œβ”€β”€ tasks.jsonl # Task descriptions
β”‚ β”‚ β”œβ”€β”€ sub_tasks.jsonl # Fine-grained sub-task annotations
β”‚ β”‚ └── camera.json # Camera parameters for all episodes
β”‚ └── videos/
β”‚ └── chunk-000/
β”‚ β”œβ”€β”€ observation.images.head_camera/
β”‚ β”‚ β”œβ”€β”€ episode_000000.mp4
β”‚ β”‚ └── ...
β”‚ β”œβ”€β”€ observation.images.left_hand_camera/
β”‚ β”‚ β”œβ”€β”€ episode_000000.mp4
β”‚ β”‚ └── ...
β”‚ └── observation.images.right_hand_camera/
β”‚ β”œβ”€β”€ episode_000000.mp4
β”‚ └── ...
β”œβ”€β”€ Arrange_the_items_on_the_conference_table/
β”‚ └── ...
β”œβ”€β”€ Clean_the_convenience_store/
β”‚ └── ...
└── ...
```
Each dataset folder contains:
- `data/` - Parquet files for observations and actions
- `meta/` - JSONL files for episodes, tasks, and annotations
- `videos/` - MP4 videos from multiple camera angles
## Tasks
The following **35 tasks** are included in this dataset:
| Category | Task List |
|----------|-----------|
| **Kitchen Tasks** | Arrange the cups, Stack the cups, Clean the kitchen counter, Tidy up the cooking counter, Cook rice using an electric rice cooker, Steam rice in a rice cooker, Steam potatoes, Steam buns, Make toast, Prepare the bread |
| **Organization Tasks** | Organize the TV cabinet, Organize the magazines, Organize the pen holder, Organize the repair tools, Organize the toys, Organize the glass tube on the rack, Replace the tissues and arrange them, Tidy up the children's room, Tidy up the conference room table, Place the books, Place the hairdryer, Place the slippers, Take down the book, Hang out the clothes to dry |
| **Convenience Store Tasks** | Clean the convenience store, Replenish tea bags |
| **Industrial Tasks** | Cable plugging, Move industrial parts to different plastic boxes, Pack the badminton shuttlecock |
| **Daily Life Tasks** | Collect the mail, Put the milk in the refrigerator, Refill the laundry detergent, Prepare the birthday cake, Take out the trash |
## Understanding the Dataset Format
This dataset follows the **LeRobot v2.1** format. Each task directory contains:
- **`data/`**: Parquet files storing time-series data (proprioceptive state, actions, timestamps)
- **`meta/`**: JSON/JSONL files containing metadata, episode information, and annotations
- **`videos/`**: MP4 video files from three camera perspectives
### Key Metadata Files
- **`meta/info.json`**: Contains dataset-level metadata, including:
- Total number of episodes, frames, videos
- Feature definitions (shapes and names of actions and observations)
- Video specifications (resolution, codec, frame rate)
- Robot type and codebase version
- **`meta/episodes.jsonl`**: Each line is a JSON object representing one episode, containing:
- `episode_index`: Episode identifier
- `length`: Number of frames in the episode
- `tasks`: List of task descriptions
- `videos`: Video file paths for each camera
- **`meta/sub_tasks.jsonl`**: Fine-grained annotations for each episode, including:
- `task_steps`: List of atomic skill segments with start/end frames
- `success_rating`: Overall task success rating (1-5 points)
- `quality_assessments`: Detailed quality metrics (pass/fail/valid)
- `notes`: Annotation metadata
- **`meta/camera.json`**: Camera intrinsics and extrinsics for each episode
## Loading and Using the Dataset
This dataset is compatible with the [LeRobot library](https://github.com/huggingface/lerobot). Here's how to load and use it:
```python
from lerobot.common.datasets.lerobot_dataset import LeRobotDataset
# Load a specific task
dataset_path = "RealSource-World/Arrange_the_cups"
repo_id = "RealSourceData/RealSource-World"
# Initialize the dataset
dataset = LeRobotDataset(dataset_path, repo_id=repo_id)
# Access episode data
episode_0 = dataset[0] # First frame of the first episode
# Episode metadata
episode_info = dataset.episode_data[0]
# Iterate through episodes
for episode_idx in range(len(dataset.episode_data)):
episode_length = dataset.episode_data[episode_idx]["length"]
print(f"Episode {episode_idx} has {episode_length} frames")
# Visualize episode
dataset.show_video(episode_idx=0, video_key="observation.images.head_camera")
```
## Data Format Details
### Data Fields
Each parquet file in `data/` contains the following columns:
| Column | Description |
|--------|-------------|
| `observation` | Multi-modal observations (images, state) |
| `action` | Robot control commands |
| `reward` | Reward signals (optional) |
| `terminated` | Episode termination flag |
| `truncated` | Truncation flag |
| `language_instruction` | Natural language task description |
### Proprioceptive State (71 Dimensions)
The `observation.state` field contains comprehensive proprioceptive information:
| Index Range | Component | Description |
|-------------|-----------|-------------|
| 0-15 | Joint Positions | 7 joints Γ— 2 arms + 2 grippers = 16 DOF |
| 16 | Lift Position | Mobile base lift height |
| 17-22 | Left Arm Force/Torque | 6D force (fx, fy, fz, mx, my, mz) |
| 23-28 | Right Arm Force/Torque | 6D force (fx, fy, fz, mx, my, mz) |
| 29-35 | Left Arm Joint Velocities | 7 joints = 7 DOF |
| 36-42 | Right Arm Joint Velocities | 7 joints = 7 DOF |
| 43-49 | Left End-Effector Pose | Position (x, y, z) + Quaternion (qw, qx, qy, qz) |
| 50-56 | Right End-Effector Pose | Position (x, y, z) + Quaternion (qw, qx, qy, qz) |
| 57-63 | Left Arm Joint Accelerations | 7 joints = 7 DOF |
| 64-70 | Right Arm Joint Accelerations | 7 joints = 7 DOF |
### State Field Names
```python
[
"LeftFollowerArm_Joint1.pos", ..., "LeftFollowerArm_Joint7.pos",
"LeftGripper.pos",
"RightFollowerArm_Joint1.pos", ..., "RightFollowerArm_Joint7.pos",
"RightGripper.pos",
"Lift.position",
"LeftForce.fx", "LeftForce.fy", "LeftForce.fz",
"LeftForce.mx", "LeftForce.my", "LeftForce.mz",
"RightForce.fx", "RightForce.fy", "RightForce.fz",
"RightForce.mx", "RightForce.my", "RightForce.mz",
"LeftJoint_Vel1", ..., "LeftJoint_Vel7",
"RightJoint_Vel1", ..., "RightJoint_Vel7",
"LeftEnd_X", "LeftEnd_Y", "LeftEnd_Z",
"LeftEnd_Qw", "LeftEnd_Qx", "LeftEnd_Qy", "LeftEnd_Qz",
"RightEnd_X", "RightEnd_Y", "RightEnd_Z",
"RightEnd_Qw", "RightEnd_Qx", "RightEnd_Qy", "RightEnd_Qz",
"LeftJoint_Acc1", ..., "LeftJoint_Acc7",
"RightJoint_Acc1", ..., "RightJoint_Acc7"
]
```
### Action Space (17 Dimensions)
The `action` field contains commands sent to the robot:
| Component | Description |
|-----------|-------------|
| 0-6 | Left arm joint positions (7 DOF) |
| 7 | Left gripper position |
| 8-14 | Right arm joint positions (7 DOF) |
| 15 | Right gripper position |
| 16 | Lift platform command |
### Action Field Names
```python
[
"LeftLeaderArm_Joint1.pos", ..., "LeftLeaderArm_Joint7.pos",
"LeftGripper.pos",
"RightLeaderArm_Joint1.pos", ..., "RightLeaderArm_Joint7.pos",
"RightGripper.pos",
"Lift.command"
]
```
### Visual Observations
Each episode contains synchronized video from three camera perspectives:
- **`observation.images.head_camera`**: Head/overhead perspective
- **`observation.images.left_hand_camera`**: Camera mounted on left end-effector
- **`observation.images.right_hand_camera`**: Camera mounted on right end-effector
**Video Specifications:**
- Resolution: 720 Γ— 1280 pixels
- Frame Rate: 30 FPS
- Codec: H.264
- Format: MP4
### Camera Parameters
Camera parameters for each episode are stored in the `meta/camera.json` file, keyed by `episode_XXXXXX`. Camera parameters include intrinsics (camera matrix and distortion coefficients) and extrinsics (hand-eye calibration parameters).
#### File Structure
The `camera.json` file contains camera parameters for all episodes:
```json
{
"episode_000000": {
"camera_ids": {
"head": "245022300889",
"left_arm": "245022301980",
"right_arm": "245022300408",
"foot": ""
},
"camera_parameters": {
"head": {
"720P": {
"MTX": [[648.57, 0, 645.54], [0, 647.80, 375.38], [0, 0, 1]],
"DIST": [-0.0513, 0.0587, -0.0006, 0.00096, -0.0186]
},
"480P": { ... }
},
"left_arm": { ... },
"right_arm": { ... }
},
"hand_eye": {
"left_arm_in_eye": {
"R": [[...], [...], [...]],
"T": [x, y, z]
},
"right_arm_in_eye": { ... },
"left_arm_to_eye": { ... },
"right_arm_to_eye": { ... }
}
},
"episode_000001": { ... }
}
```
#### Camera Intrinsics
Each camera (head, left_arm, right_arm) contains intrinsics for two resolutions:
- **`MTX`**: 3Γ—3 camera intrinsic matrix
```
[fx 0 cx]
[0 fy cy]
[0 0 1]
```
- `fx`, `fy`: Focal length (in pixels)
- `cx`, `cy`: Principal point (optical center) coordinates (in pixels)
- **`DIST`**: 5-element distortion coefficients (k1, k2, p1, p2, k3)
- Used for correcting radial and tangential distortion
**Available Resolutions:**
- `720P`: 720p video parameters (720 Γ— 1280)
- `480P`: 480p video parameters (480 Γ— 640)
#### Hand-Eye Calibration (Extrinsics)
The `hand_eye` section contains transformation relationships between robot end-effectors and cameras:
- **`left_arm_in_eye`**: Transformation from left wrist camera (left end-effector camera) to left manipulator end center
- `R`: 3Γ—3 rotation matrix
- `T`: 3Γ—1 translation vector [x, y, z] (in meters)
- Represents the position and orientation of the left wrist-mounted camera relative to the left manipulator end center
- **`right_arm_in_eye`**: Transformation from right wrist camera (right end-effector camera) to right manipulator end center
- Represents the position and orientation of the right wrist-mounted camera relative to the right manipulator end center
- **`left_arm_to_eye`**: Transformation from head camera to left manipulator base coordinate system
- `R`: 3Γ—3 rotation matrix
- `T`: 3Γ—1 translation vector [x, y, z] (in meters)
- Represents the position and orientation of the head camera relative to the left manipulator base coordinate system
- **`right_arm_to_eye`**: Transformation from head camera to right manipulator base coordinate system
- Represents the position and orientation of the head camera relative to the right manipulator base coordinate system
These parameters support the following coordinate transformations:
- Conversion between robot end-effector poses and camera image coordinates
- Conversion between 3D positions in robot space and image pixel coordinates
- Multi-view geometry operations and calibration
- Conversion between wrist camera coordinate system and end center
- Conversion between head camera coordinate system and manipulator base coordinate system
#### Camera IDs
Each camera has a unique identifier:
- **`head`**: Head camera ID
- **`left_arm`**: Left end-effector camera ID
- **`right_arm`**: Right end-effector camera ID
- **`foot`**: Foot camera ID (if available)
### Sub-task Annotations
Each episode in `meta/sub_tasks.jsonl` contains detailed annotations:
```json
{
"task": "Separate the two stacked cups in the dish and place them on the two sides of the dish.",
"language": "en",
"task_index": 0,
"episode_index": 0,
"task_steps": [
{
"step_name": "Left arm picks up the stack of cups from the center of the plate",
"start_frame": 100,
"end_frame": 180,
"description": "Left arm picks up the stack of cups from the center of the plate",
"duration_frames": 80
},
...
],
"success_rating": 5,
"notes": "annotation_date: 2025/11/13",
"quality_assessments": {
"overall_valid": "VALID",
"movement_fluency": "PASS",
"grasp_success": "PASS",
"placement_quality": "PASS",
...
},
"total_frames": 946
}
```
#### Quality Assessment Metrics
- **`overall_valid`**: Overall episode validity (valid/invalid)
- **`movement_fluency`**: Robot movement fluency (pass/fail)
- **`grasp_success`**: Grasp action success (pass/fail)
- **`placement_quality`**: Object placement quality (pass/fail)
- **`no_drop`**: No objects dropped during task (pass/fail)
- **`grasp_collisions`**: No collisions during grasping (pass/fail)
- **`arm_collisions`**: No arm collisions (pass/fail)
- **`operation_completeness`**: Task completion status (pass/fail)
- And other metrics...
## Dataset Statistics
### Overall Statistics
- **Total Tasks**: 35
- **Total Episodes**: 11,428
- **Total Frames**: 14,085,107
- **Total Videos**: 34,284 (3 cameras Γ— 11,428 episodes)
- **Robot Type**: RS-02 (dual-arm humanoid robot)
- **Dataset Format**: LeRobot v2.1
- **Video Resolution**: 720 Γ— 1280
- **Frame Rate**: 30 FPS
## Task Distribution
The dataset contains diverse manipulation tasks across multiple domains:
| Category | Task List |
|----------|-----------|
| **Kitchen Tasks** | Arrange the cups, Stack the cups, Clean the kitchen counter, Tidy up the cooking counter, Cook rice using an electric rice cooker, Steam rice in a rice cooker, Steam potatoes, Steam buns, Make toast, Prepare the bread |
| **Organization Tasks** | Organize the TV cabinet, Organize the magazines, Organize the pen holder, Organize the repair tools, Organize the toys, Organize the glass tube on the rack, Replace the tissues and arrange them, Tidy up the children's room, Tidy up the conference room table, Place the books, Place the hairdryer, Place the slippers, Take down the book, Hang out the clothes to dry |
| **Convenience Store Tasks** | Clean the convenience store, Replenish tea bags |
| **Industrial Tasks** | Cable plugging, Move industrial parts to different plastic boxes, Pack the badminton shuttlecock |
| **Daily Life Tasks** | Collect the mail, Put the milk in the refrigerator, Refill the laundry detergent, Prepare the birthday cake, Take out the trash |
## Robot URDF Model
RealSource World dataset is collected using the **RS-02** dual-arm humanoid robot. For simulation, visualization, and research purposes, we provide the URDF (Unified Robot Description Format) model of the RS-02 robot.
### RS-02 Robot Specifications
- **Robot Type**: Dual-arm humanoid robot
- **Total Links**: 46 links
- **Total Joints**: 45 joints
- **Manipulator Arms**: 2 Γ— 7-DOF manipulator arms (left and right)
- **End Effectors**: Dual grippers, each with 8 DOF
- **Base**: Mobile base with wheels and lift platform
- **Sensors**: Head camera, left hand camera, right hand camera
### URDF Package Structure
The RS-02 URDF package contains:
```
RS-02/
β”œβ”€β”€ urdf/
β”‚ β”œβ”€β”€ RS-02.urdf # Main URDF file (59KB)
β”‚ └── RS-02.csv # Joint configuration data
β”œβ”€β”€ meshes/ # 3D mesh models (46 STL files)
β”‚ β”œβ”€β”€ base_link.STL
β”‚ β”œβ”€β”€ L_Link_1-7.STL # Left arm links
β”‚ β”œβ”€β”€ R_Link_1-7.STL # Right arm links
β”‚ β”œβ”€β”€ ltool_*.STL # Left gripper components
β”‚ β”œβ”€β”€ rtool_*.STL # Right gripper components
β”‚ β”œβ”€β”€ head_*.STL # Head components
β”‚ └── camera_*.STL # Camera mounts
β”œβ”€β”€ config/
β”‚ └── joint_names_RS-02.yaml # Joint name configuration
β”œβ”€β”€ launch/
β”‚ β”œβ”€β”€ display.launch # RViz visualization
β”‚ └── gazebo.launch # Gazebo simulation
└── package.xml # ROS package metadata
```
### Using the URDF Model
#### ROS/ROS2 Users
The URDF model can be used with ROS tools:
**Visualize in RViz:**
```bash
roslaunch RS-02 display.launch
```
**Simulate in Gazebo:**
```bash
roslaunch RS-02 gazebo.launch
```
## License and Citation
This dataset is released under the **CC BY-NC-SA 4.0** license (Attribution-NonCommercial-ShareAlike).
### License
This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-sa/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA.
### Citation
If you use this dataset in your research, please cite:
```bibtex
@misc{realsourceworld,
title={RealSource World: A Large-Scale Real-World Dual-Arm Manipulation Dataset},
author={RealSource},
howpublished={\url{https://huggingface.co/datasets/RealSourceData/RealSource-World}},
year={2025}
}
```