| --- |
| license: apache-2.0 |
| tags: |
| - robotics |
| - vla-model |
| - embodied-ai |
| - task-generation |
| - humanoid-robots |
| - lerobot |
| - pytorch |
| - agentic-framework |
| datasets: |
| - robogene |
| --- |
| |
|
|
| # Dataset Card for RoboGene-Dataset |
|
|
| <p align="center"> |
| <a href="https://arxiv.org/abs/2602.16444"><img src="https://img.shields.io/badge/arXiv-2602.16444-B31B1B.svg" alt="Paper"></a> |
| <a href="https://huggingface.co/datasets/X-Humanoid/RoboGene"><img src="https://img.shields.io/badge/%F0%9F%A4%97%20Dataset-HuggingFace-yellow" alt="Dataset"></a> |
| <a href="LICENSE"><img src="https://img.shields.io/badge/License-Apache--2.0-blue.svg" alt="License"></a> |
| </p> |
| <p align="center"> |
| <img src="teaser.png" width="800"> |
| </p> |
| <p align="center"><em>Figure 1: Overview of the RoboGene dataset.</em></p> |
|
|
| **RoboGene** is a large-scale robotic task dataset generated by a **diversity-driven agentic framework**. Specifically designed for **Vision-Language-Action (VLA)** model pre-training, it addresses the critical bottlenecks of limited scene variety and insufficient physical grounding in existing datasets, significantly enhancing the generalization capabilities of humanoid robots in complex real-world tasks. |
|
|
|
|
|
|
| ## Dataset Details |
|
|
| ### Dataset Description |
|
|
| RoboGene overcomes the limitations of traditional human teleoperation for data collection. By leveraging foundation models (e.g., GPT-4o) as generators within the **RoboGene Agent Loop**—which integrates retrieval, generation, and self-reflection—the framework ensures that every generated instruction is not only semantically diverse but also physically executable in structured environments. |
|
|
| - **Curated by:** Beijing Innovation Center of Humanoid Robotics, Peking University, Beihang University, Beijing Institute of Technology. |
| - **Shared by:** RoboGene Project Team |
| - **Language(s) (NLP):** Chinese (Task Instructions) |
| - **License:** Apache-2.0 |
|
|
| ### Dataset Sources |
|
|
| - **Demo:** (https://robogene-boost-vla.github.io/) |
| - **Paper:** *RoboGene: Boosting VLA Pre-training via Diversity-Driven Agentic Framework for Real-World Task Generation* |
| - **Framework Compatibility:** Native support for the [LeRobot](https://github.com/huggingface/lerobot) framework. |
|
|
| ## Uses |
|
|
| ### Direct Use |
|
|
| This dataset serves as a **pre-training foundation** for VLA models: |
| - **Massive-Scale Pre-training:** Enhances model robustness against novel objects, lighting variations, and background clutter. |
| - **Zero/Few-shot Adaptation:** Experiments show that models pre-trained on RoboGene require only 50 human demonstrations to quickly adapt to complex downstream tasks. |
| - **Dual-arm Coordination Research:** Contains high-quality trajectories for sophisticated dual-arm manipulation. |
|
|
| ### Out-of-Scope Use |
|
|
| - Deploying pre-trained policies directly to heavy industrial equipment without safety validation or simulation-based verification is strictly prohibited. |
|
|
| ## Dataset Structure |
|
|
| The dataset is organized by scenes with a balanced distribution. Each episode includes: |
| - **Observations:** Multi-view RGB images (front-view, wrist-view) and proprioceptive states. |
| - **Actions:** High-frequency control sequences for joints or end-effectors. |
| - **Instructions:** Verified, high-quality natural language task descriptions. |
|
|
| ### Task Highlights |
| - **100+ Diverse Skills:** Ranging from basic pick-and-place to complex multi-stage assembly and tool usage. |
| - **Rich Scenarios:** Comprehensive coverage of Kitchen, Laboratory, Medical, Office, Retail, Education, Industrial, and Daily Life environments. |
| - **Complex Object Spectrum:** Includes high-reflectivity materials (e.g., glass), deformable/flexible objects (e.g., cloth, ropes), and articulated/hinged mechanisms (e.g., doors, drawers, cabinets). |
|
|
|
|
|
|
| ## Dataset Creation |
|
|
| ### Curation Rationale |
|
|
| While human-curated data is physically valid, it often lacks structural diversity. Conversely, traditional LLM-generated tasks frequently fail to align with physical constraints. RoboGene bridges this gap by embedding physical grounding into an automated generation loop, resulting in pre-training performance that significantly outperforms both purely human-curated and baseline foundation-model datasets. |
|
|
|
|
| ## Bias, Risks, and Limitations |
|
|
| - **Sim-to-Real Gap:** Although verified for physical grounding, minor sim-to-real discrepancies may still occur in extremely high-precision contact mechanics tasks. |
| - **Domain Scope:** While covering 8+ major scenarios, the dataset may not account for highly specialized or hazardous industrial environments not represented in the training distribution. |
|
|
| ### Performance Benchmarks |
| Empirical results demonstrate that the RoboGene pre-training policy significantly outperforms human-curated baselines across five unseen dual-arm tasks. Most notably, in the complex DFR-BuildBlocks scenario, RoboGene achieves a success rate five times higher than that of existing foundation models (25% vs. 5%). Furthermore, our approach exhibits superior generalization capabilities, maintaining robust performance in novel environments where baseline policies frequently fail to adapt. |
| ## structure of the dataset |
|
|
|
|
| Each task follows the LeRobot format: |
| ``` |
| task_name/ |
| ├── data/ |
| │ └── chunk-000/ |
| │ ├── episode_000000.parquet |
| │ ├── episode_000001.parquet |
| │ └── ... |
| ├── meta/ |
| │ ├── info.json # |
| │ ├── tasks.jsonl # task description |
| │ ├── episodes.jsonl # |
| │ └── episodes_stats.jsonl |
| └── videos/ |
| └── chunk-000/ |
| ├── observation.rgb_images.camera_front/ |
| ├── observation.rgb_images.camera_left/ |
| ├── observation.rgb_images.camera_right/ |
| ├── observation.rgb_images.camera_top/ |
| ├── observation.rgb_images.camera_wrist_left/ |
| └── observation.rgb_images.camera_wrist_right/ |
| ``` |
|
|
|
|
|
|
| ### camera configuration(six RGB cameras) |
| - `camera_front` |
| - `camera_left` |
| - `camera_right` |
| - `camera_top` |
| - `camera_wrist_left` |
| - `camera_wrist_right` |
| ## Citation |
|
|
| If you use this dataset in your research, please cite: |
|
|
| **BibTeX:** |
|
|
| ```latex |
| @article{zhang2026robogene, |
| title={RoboGene: Boosting VLA Pre-training via Diversity-Driven Agentic Framework for Real-World Task Generation}, |
| author={Zhang, Yixue and Wu, Kun and Gao, Zhi and Zhao, Zhen and Ren, Pei and Xu, Zhiyuan and Liao, Fei and Wang, Xinhua and Fan, Shichao and Wu, Di and others}, |
| journal={arXiv preprint arXiv:2602.16444}, |
| year={2026} |
| } |