File size: 17,359 Bytes
19957bc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19f636f
19957bc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6b51c91
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19957bc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
---
license: mit
---
<p align="center">
  <h2>Dexora: Open-Source VLA for High-DoF Bimanual Dexterity</h2>
</p>

<p align="center">
  <a href="#"><img src="https://img.shields.io/badge/arXiv-2026.xxxxx-B31B1B.svg" alt="arXiv"></a>
  <a href="https://github.com/ZZongzheng0918/Dexora?tab=readme-ov-file"><img src="https://img.shields.io/badge/Project-Page-blue.svg" alt="Project Page"></a>
  <a href="LICENSE"><img src="https://img.shields.io/badge/License-MIT-green.svg" alt="License"></a>
</p>


## πŸ”₯ News & Updates


- **2025-12-03**: Released the full **Real-World Dataset** (**10K episodes**)  on Hugging Face.

--

## πŸ“Š Dataset Overview

The Dexora corpus combines **high-fidelity real-world teleoperation data** with a **large-scale simulated dataset** designed to match the embodiment of the physical robot.

### A. Dexora Real-World Dataset (High-Fidelity)

The Dexora real-world dataset consists of **11.5K teleoperated episodes**, **2.92M frames**, and **40.5 hours** of data. Demonstrations are collected using a **hybrid teleoperation system** that couples an **Exoskeleton** (for arm control) with **Vision Pro** (for dexterous hand control), enabling precise 36-DoF bimanual manipulation on real hardware.

<p align="center">
  <img src="assets/image/dataset.gif" alt="Dexora Multi-view Dataset" width="100%">
</p>

<p align="center">
  <i>Video 1. <b>Synchronized Multi-View Recordings.</b> High-resolution streams from ego-centric, third-person, and wrist-mounted cameras, synchronized with 36-DoF robot proprioception.</i>
</p>

<p align="center">
  <img src="assets/image/real-data.JPG" alt="Dexora Real-World Dataset Mosaic" width="100%">
</p>

<p align="center">
  <i>Fig 1. <b>High-Fidelity Real-World Scenes.</b> Collected via our hybrid teleoperation system (Exoskeleton for arm + Vision Pro for hand), this dataset covers <b>347 objects</b> across diverse environments. It captures varying lighting conditions, background clutter, and precise bimanual interactions essential for robust policy learning. Panels (a–d) correspond to four task categories: <b>pick-and-place</b>, <b>assembly</b>, <b>articulation</b>, and <b>dexterous manipulation</b>.</i>
</p>

<p align="center">
  <img src="assets/image/Categorized%20Robot%20Task%20Trajectory%20Distribution.png" alt="Dexora Task Categories" width="120%">
</p>

<p align="center">
  <img src="assets/image/Robot%20Arm%20Task%20Trajectory%20Distribution.png" alt="Dexora Robot Arm Trajectory Distribution" width="120%">
</p>

<p align="center">
  <i>Fig 2. <b>Task Categories & Action Distribution.</b> Unlike standard gripper datasets, Dexora emphasizes high-DoF dexterity. The real-world data distribution includes <b>Dexterous Manipulation (20%)</b> (e.g., <i>Twist Cap</i>, <i>Use Pen</i>, <i>Cut Leek</i>) and <b>Assembly (15%)</b> (e.g., <i>Separate Nested Bowls</i>, <i>Stack Ring Blocks</i>), in addition to <b>Articulated Objects (10%)</b> and <b>Pick-and-Place (55%)</b>.</i>
</p>

Both the episodes and annotations follow the **LIBERO-2.1 standard**, including synchronized **RGB observations**, **robot proprioception**, **actions**, and **language instructions**.

### Object Inventory & Reproducibility

Reproducibility is a **core value** of Dexora. To enable other labs and industry teams to **faithfully recreate** our environments, we release a **curated object inventory** that mirrors the physical setup used in our real-world experiments.

- **Scale**: **347 objects** across **17 semantic categories** (e.g., tools, containers, articulated objects, deformables, daily-use items).
- **Coverage**: Objects are chosen to stress **dexterous control**, **bimanual coordination**, and **long-horizon manipulation**.
- **Procurement**: Every item is linked to **Taobao** and/or **Amazon**, so researchers can rebuild the setup with minimal effort.

<p align="center">
  <a href="https://docs.google.com/spreadsheets/d/1L2cgqvIukVziXc0OwpqNkb5j8c3bzC_K/edit?usp=sharing">
    <b>πŸ“‘ Access Dexora Real-world Item List (Google Sheet)</b>
  </a>
</p>

### Inventory Metadata Schema

The released Google Sheet follows the schema below:

| Column                     | Description                                                                 |
| :------------------------- | :-------------------------------------------------------------------------- |
| **Object Name (EN & CN)** | Bilingual identification for global researchers.                            |
| **Task Type**             | One of: `pick-and-place`, `assemble`, `articulation`, `dexterous`.          |
| **Purchase Link**         | Direct links to **Taobao** & **Amazon** for easy procurement and restock.   |

You can **filter by task type**, **category**, or **store** to design controlled benchmarks or new task suites on top of Dexora.


### B. Dexora Simulation Dataset (Large-Scale)

The Dexora simulation dataset contains **100K episodes** generated in **MuJoCo**, using the same **36-DoF dual-arm, dual-hand** embodiment as the real robot. It provides large-scale, embodiment-matched experience focused on core skills such as **pick-and-place**, **assembly**, and **articulation**, which can be used for pre-training basic competence before **fine-tuning on the real-world dataset**.

### Summary Statistics (Sim vs Real)

| **Split**        | **Episodes** | **Frames** | **Hours (approx.)** | **Task Types**                                                                 |
| :--------------- | -----------: | ---------: | -------------------: | :----------------------------------------------------------------------------- |
| **Simulated**    | **100K**     | **6.5M**   | TBD                  | Pick-and-place, assembly, articulation |
| **Real-World**   | **10K**      | **3.2M**   | **177.5**            | Teleoperated bimanual tasks with high-DoF hands, cluttered scenes, fine-grain object interactions |


## πŸ“‚ Data Structure

Dexora follows the **LIBERO-2.1** dataset standard. Each episode is stored as a self-contained trajectory with:

- **Observations**: multi-view RGB (and optionally depth), segmentation masks (when available).
- **Robot State**: joint positions/velocities for dual arms and dual hands, gripper/hand states.
- **Actions**: low-level control commands compatible with 36-DoF bimanual control.
- **Language**: High-level task descriptions. We provide **5 diverse natural language instructions** per task, distributed evenly across all trajectories to enhance linguistic diversity.

An example high-level directory layout is:

```text
data
β”œβ”€β”€ real
β”‚   β”œβ”€β”€ articulation
β”‚   β”‚   └── ...
β”‚   β”œβ”€β”€ assembly
β”‚   β”‚   └── ...
β”‚   β”œβ”€β”€ dexterous manipulation                     
β”‚   β”‚   β”œβ”€β”€ data 
β”‚   β”‚   β”‚   β”œβ”€β”€ chunk-000 
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ episode_000000.parquet
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ episode_000001.parquet
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ episode_000002.parquet
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”‚   β”‚   β”œβ”€β”€ chunk-001
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”‚   β”œβ”€β”€ meta
β”‚   β”‚   β”‚   β”œβ”€β”€ episodes.jsonl  
β”‚   β”‚   β”‚   β”œβ”€β”€ episodes_stats.jsonl   
β”‚   β”‚   β”‚   β”œβ”€β”€ info.json      
β”‚   β”‚   β”‚   β”œβ”€β”€ modality.json  
β”‚   β”‚   β”‚   β”œβ”€β”€ stats.json     
β”‚   β”‚   β”‚   β”œβ”€β”€ tasks.jsonl   
β”‚   β”‚   β”œβ”€β”€ videos
β”‚   β”‚   β”‚   β”œβ”€β”€ chunk-000 
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ observation.images.front
β”‚   β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ episode_000000.mp4
β”‚   β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ episode_000001.mp4
β”‚   β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”‚   β”‚   β”œβ”€β”€ chunk-001
β”‚   β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”‚   β”‚   β”œβ”€β”€ ...
β”‚   β”œβ”€β”€ pick_and_place
β”‚   β”‚   └── ...
β”‚   β”œβ”€β”€ ...
β”œβ”€β”€ sim
β”‚   β”œβ”€β”€ ...

```

> **Note**: The exact folder names and file formats may be updated as we finalize the public release, but the overall **episode-centric LIBERO-2.1 structure** will be preserved.

---

### [meta/info.json](meta/info.json):
```json
{
    "codebase_version": "v2.1",
    "robot_type": "airbot_play",
    "total_episodes": 11517,
    "total_frames": 2919110,
    "total_tasks": 201,
    "total_videos": 46068,
    "total_chunks": 12,
    "chunks_size": 1000,
    "fps": 20,
    "splits": {
        "train": "0:2261"
    },
    "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet",
    "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4",
    "features": {
        "observation.images.top": {
            "dtype": "video",
            "shape": [
                480,
                640,
                3
            ],
            "names": [
                "height",
                "width",
                "channels"
            ],
            "info": {
                "video.height": 480,
                "video.width": 640,
                "video.codec": "av1",
                "video.pix_fmt": "yuv420p",
                "video.is_depth_map": false,
                "video.fps": 20,
                "video.channels": 3,
                "has_audio": false
            }
        },
        "observation.images.wrist_left": {
            "dtype": "video",
            "shape": [
                480,
                640,
                3
            ],
            "names": [
                "height",
                "width",
                "channels"
            ],
            "info": {
                "video.height": 480,
                "video.width": 640,
                "video.codec": "av1",
                "video.pix_fmt": "yuv420p",
                "video.is_depth_map": false,
                "video.fps": 20,
                "video.channels": 3,
                "has_audio": false
            }
        },
        "observation.images.wrist_right": {
            "dtype": "video",
            "shape": [
                480,
                640,
                3
            ],
            "names": [
                "height",
                "width",
                "channels"
            ],
            "info": {
                "video.height": 480,
                "video.width": 640,
                "video.codec": "av1",
                "video.pix_fmt": "yuv420p",
                "video.is_depth_map": false,
                "video.fps": 20,
                "video.channels": 3,
                "has_audio": false
            }
        },
        "observation.images.front": {
            "dtype": "video",
            "shape": [
                480,
                640,
                3
            ],
            "names": [
                "height",
                "width",
                "channels"
            ],
            "info": {
                "video.height": 480,
                "video.width": 640,
                "video.codec": "av1",
                "video.pix_fmt": "yuv420p",
                "video.is_depth_map": false,
                "video.fps": 20,
                "video.channels": 3,
                "has_audio": false
            }
        },
        "observation.state": {
            "dtype": "float32",
            "shape": [
                39
            ],
            "names": [
                "left_arm_joint_1",
                "left_arm_joint_2",
                "left_arm_joint_3",
                "left_arm_joint_4",
                "left_arm_joint_5",
                "left_arm_joint_6",
                "right_arm_joint_1",
                "right_arm_joint_2",
                "right_arm_joint_3",
                "right_arm_joint_4",
                "right_arm_joint_5",
                "right_arm_joint_6",
                "left_hand_joint_1",
                "left_hand_joint_2",
                "left_hand_joint_3",
                "left_hand_joint_4",
                "left_hand_joint_5",
                "left_hand_joint_6",
                "left_hand_joint_7",
                "left_hand_joint_8",
                "left_hand_joint_9",
                "left_hand_joint_10",
                "left_hand_joint_11",
                "left_hand_joint_12",
                "right_hand_joint_1",
                "right_hand_joint_2",
                "right_hand_joint_3",
                "right_hand_joint_4",
                "right_hand_joint_5",
                "right_hand_joint_6",
                "right_hand_joint_7",
                "right_hand_joint_8",
                "right_hand_joint_9",
                "right_hand_joint_10",
                "right_hand_joint_11",
                "right_hand_joint_12",
                "head_joint_1",
                "head_joint_2",
                "spine_joint"
            ]
        },
        "action": {
            "dtype": "float32",
            "shape": [
                39
            ],
            "names": [
                "left_arm_joint_1",
                "left_arm_joint_2",
                "left_arm_joint_3",
                "left_arm_joint_4",
                "left_arm_joint_5",
                "left_arm_joint_6",
                "right_arm_joint_1",
                "right_arm_joint_2",
                "right_arm_joint_3",
                "right_arm_joint_4",
                "right_arm_joint_5",
                "right_arm_joint_6",
                "left_hand_joint_1",
                "left_hand_joint_2",
                "left_hand_joint_3",
                "left_hand_joint_4",
                "left_hand_joint_5",
                "left_hand_joint_6",
                "left_hand_joint_7",
                "left_hand_joint_8",
                "left_hand_joint_9",
                "left_hand_joint_10",
                "left_hand_joint_11",
                "left_hand_joint_12",
                "right_hand_joint_1",
                "right_hand_joint_2",
                "right_hand_joint_3",
                "right_hand_joint_4",
                "right_hand_joint_5",
                "right_hand_joint_6",
                "right_hand_joint_7",
                "right_hand_joint_8",
                "right_hand_joint_9",
                "right_hand_joint_10",
                "right_hand_joint_11",
                "right_hand_joint_12",
                "head_joint_1",
                "head_joint_2",
                "spine_joint"
            ]
        },
        "timestamp": {
            "dtype": "float32",
            "shape": [
                1
            ],
            "names": null
        },
        "frame_index": {
            "dtype": "int64",
            "shape": [
                1
            ],
            "names": null
        },
        "episode_index": {
            "dtype": "int64",
            "shape": [
                1
            ],
            "names": null
        },
        "index": {
            "dtype": "int64",
            "shape": [
                1
            ],
            "names": null
        },
        "task_index": {
            "dtype": "int64",
            "shape": [
                1
            ],
            "names": null
        }
    }
}
```

## πŸ“₯ Usage

### 1. Environment Setup

We recommend using **conda** to manage dependencies:

```bash
conda create -n dexora python=3.10 -y
conda activate dexora

# Clone this repository
git clone <your-dexora-repo-url>.git
cd Dexora

# Install Python dependencies (example)
pip install -r requirements.txt
```

If you plan to train or fine-tune VLAs, please ensure that **PyTorch**, **CUDA**, and any required simulation backends (e.g., Isaac, Mujoco) are installed according to your hardware.

### 2. Downloading the Dataset

- **Simulated Dexora**: Download links will be provided on the **project page** (see badge above).
- **Real-World Dexora**: High-resolution teleoperation data (RGB, proprio, actions) will be hosted via a public storage service (e.g., academic server / cloud bucket).

Typical usage:

```bash
# Example directory where you store data
export DEXORA_DATA=/path/to/dexora

# (Optional) Symlink data into this repo
ln -s $DEXORA_DATA data
```

### 3. Loading Episodes (Example)

Below is a minimal Python snippet illustrating how to load a Parquet episode from the real-world dataset:

```python
import pandas as pd
from pathlib import Path

# Example: Loading a Parquet episode from the real-world dataset
root = Path("data/real/dexterous_manipulation/data/chunk-000")
episode_path = root / "episode_000000.parquet"

# Load trajectory using pandas
df = pd.read_parquet(episode_path)

# Access data columns (Observation, Action, Proprioception)
# Note: Columns are typically flattened in Parquet format
print("Available keys:", df.columns)
print("Actions shape:", df["action"].shape)  # Example access
print("Language Instruction:", df["language_instruction"][0])
```

---

## πŸ“œ Citation

If you find Dexora useful in your research, please consider citing our paper:

```bibtex
@misc{dexora2026,
  title         = {Dexora: Open-Source VLA for High-DoF Bimanual Dexterity},
  author        = {Dexora Team},
  year          = {2026},
  archivePrefix = {arXiv},
  eprint        = {xxxx.xxxxx},
  primaryClass  = {cs.RO}
}
```

---

For questions, collaborations, or feedback, please feel free to open an issue or contact the maintainers via the project page.