File size: 6,239 Bytes
c605285
 
d2e7033
 
 
c605285
 
 
 
 
 
d2e7033
c605285
 
 
 
d2e7033
 
c605285
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d2e7033
c605285
 
 
 
 
 
 
 
 
 
 
d2e7033
c605285
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d2e7033
c605285
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d2e7033
 
c605285
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d2e7033
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
---
license: cdla-permissive-2.0
pretty_name: MOSAIC Dataset
task_categories:
- robotics
tags:
- humanoid
- motion-tracking
- teleoperation
- reinforcement-learning
- motion-retargeting
arxiv: 2602.08594
---

# MOSAIC Dataset

[**Project Page**](https://baai-humanoid.github.io/MOSAIC/) | [**Paper**](https://arxiv.org/abs/2602.08594) | [**Code**](https://github.com/BAAI-Humanoid/MOSAIC)

This repository releases the built-in **MOSAIC** multi-source motion dataset in the following paper:

**[MOSAIC: Bridging the Sim-to-Real Gap in Generalist Humanoid Motion Tracking and Teleoperation with Rapid Residual Adaptation](https://arxiv.org/abs/2602.08594)**  

The dataset is organized into:
- **Human motions** stored in an **[AMASS-style format](https://amass.is.tue.mpg.de/)**
- **Unitree G1 motions** retargeted from human motions and converted to **NPZ** for training/visualization

It includes motions from:
1. **Optical MoCap (Vicon)**
2. **Inertial MoCap (IO-AI)**
3. **Generated motions from GENMO**

Additionally, we provide a small amount of **Stage-II adaptor training data**:
- ~30 min **PICO VR** data
- ~30 min **Noitom** inertial mocap data

We also release the **GENMO prompts** used for motion generation, and the same prompts were used as guidance during **optical MoCap capture**.

---

## Data Formats & Processing Pipeline

### 1) Human motions (AMASS-style)
Human motion files follow an AMASS-style parameterization/layout for convenient reuse with common SMPL/AMASS tooling.

### 2) G1 motions (retargeted + converted to NPZ)
We retarget human motions to **Unitree G1** using:
- **[GMR](https://github.com/YanjieZe/GMR)** for retargeting (human → humanoid robot)
- **[BeyondMimic](https://github.com/HybridRobotics/whole_body_tracking)** for converting to NPZ and visualization

Concretely:
`Human (AMASS-style) → GMR retargeting → BeyondMimic conversion → *.npz`

---

## Folder Structure

<details>
<summary><b>Click to expand tree</b></summary>

```text
MOSAIC_Dataset/
├── human/
│   ├── optical_mocap/
│   │   ├── human_DA*_stageii.npz
│   │   └── ...
│   ├── inertial_mocap/
│   │   ├── mocap_data/                  # raw IO-AI recordings
│   │   │   ├── *.mcap                   # recorded streams (incl. TF)
│   │   │   └── ...
│   │   ├── amass_data/                  # converted AMASS-style files
│   │   │   ├── *.npz
│   │   │   └── ...
│   │   └── mocap_to_amass.py            # helper script for conversion
│   └── generated_genmo/
│       ├── walk/
│       │   ├── 000/
│       │   │   └── hmr4d_results.pt
│       │   └── ...
│       ├── dance/
│       ├── daily_activaties/
│       └── ...

├── G1/
│   ├── optical_mocap/
│   │   ├── g1_DA*_stageii.npz
│   │   └── ...
│   ├── inertial_mocap/
│   │   ├── g1_io_*.npz
│   │   └── ...
│   ├── generated_genmo/
│   │   ├── g1_genmo_*.npz
│   │   └── ...
│   └── adaptor_data/
│       ├── pico_VR/
│       │   ├── mocap_motion_*_smoothed.npz
│       │   └── ...
│       └── noitom_inertial_mocap/
│           ├── mocap_motion_*_smoothed_*.npz
│           └── ...

└── motion_prompt/
    ├── motion_prompt_*_.pdf
    └── ...
```
</details>

---

## How to Download

### Option A: Download the whole dataset (recommended)
  ```python
    from huggingface_hub import snapshot_download

    local_dir = snapshot_download(
        repo_id="BAAI-Humanoid/MOSAIC_Dataset",
        repo_type="dataset",
    )
    print("Downloaded to:", local_dir)
  ```

### Option B: Download a subset (recommended for large datasets)
Example: only download prompts and one subset folder.
  ```python
    from huggingface_hub import snapshot_download

    local_dir = snapshot_download(
        repo_id="BAAI-Humanoid/MOSAIC_Dataset",
        repo_type="dataset",
        allow_patterns=[
            "motion_prompt/*",
            "G1/adaptor_data/*",
            # add more patterns as needed, e.g.:
            # "human/optical_mocap/*",
            # "G1/optical_mocap/*",
        ],
    )
    print("Downloaded to:", local_dir)
  ```
### Option C: CLI download
  ```bash
    huggingface-cli download BAAI-Humanoid/MOSAIC_Dataset \
    --repo-type dataset \
    --local-dir MOSAIC_Dataset
  ```

---

## Visualization
You can visualize the G1 NPZ motions using:

- our [MOSAIC training repository](https://github.com/BAAI-Humanoid/MOSAIC), or

- [BeyondMimic](https://github.com/HybridRobotics/whole_body_tracking)

Tip: If you only want to inspect a few sequences quickly, download a small subset via allow_patterns and visualize those NPZ files first.

---

## Citation
If you use this dataset, please cite our paper:
```bibtex
@article{sun2026mosaic,
  title   = {MOSAIC: Bridging the Sim-to-Real Gap in Generalist Humanoid Motion Tracking and Teleoperation with Rapid Residual Adaptation},
  author  = {Zhenguo Sun and Bo-Sheng Huang and Yibo Peng and Xukun Li and Jingyu Ma and Yu Sun and Zhe Li and Haojun Jiang and Biao Gao and Zhenshan Bing and Xinlong Wang and Alois Knoll},
  journal = {arXiv preprint arXiv:2602.08594},
  year    = {2026}
}
```

---

## License
This dataset is released under CDLA-Permissive-2.0.

**Important**: External tools/formats referenced in this dataset card (AMASS schema, GENMO, GMR, BeyondMimic, IO-AI tooling) each have their own licenses/terms—please follow the original terms when using those resources.

---

## Acknowledgements / References
- [AMASS](https://amass.is.tue.mpg.de/)
- [GMR retargeting](https://github.com/YanjieZe/GMR)
- [BeyondMimic](https://github.com/HybridRobotics/whole_body_tracking)
- [GENMO](https://github.com/NVlabs/GENMO)
- [IO-AI mocap tools](https://github.com/ioai-tech/io_dev_tools_ros2/tree/master/io_mocap)

---

## Contact / Issues
Please open an issue in the relevant code repository or reach out via your project contact channels if you find problems with the dataset structure, conversion scripts, or visualization.