File size: 4,720 Bytes
cfa455d
 
367b8de
 
 
62a83f5
 
367b8de
 
 
 
 
 
 
62a83f5
367b8de
62a83f5
cfa455d
 
367b8de
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
 
 
 
 
 
 
cfa455d
367b8de
cfa455d
367b8de
fb2eaa6
367b8de
 
 
 
fb2eaa6
367b8de
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
 
 
 
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
 
 
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
 
 
cfa455d
367b8de
cfa455d
367b8de
 
 
cfa455d
367b8de
cfa455d
367b8de
 
 
cfa455d
367b8de
cfa455d
367b8de
 
 
cfa455d
367b8de
cfa455d
367b8de
cfa455d
367b8de
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7f6dd45
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
---
license: odbl
task_categories:
  - robotics
  - video-classification
  - image-classification
  - object-detection
tags:
  - dexterous-manipulation
  - hand-object-interaction
  - motion-capture
  - physics-simulation
  - rgbd
  - contact-forces
  - computer-vision
size_categories:
  - 10K<n<100K
---

# DexCanvas: Dexterous Manipulation Dataset v0.1

**⚠️ TEST RELEASE**: This is a preview version containing 1% of the full dataset. Contact force data is not included in v0.1.

DexCanvas is a large-scale hybrid dataset for robotic hand-object interaction research, combining real human demonstrations with physics-validated simulation data.

## Dataset Statistics (v0.1 Test Release)

- **Total Frames**: ~30 million multi-view RGB-D frames
- **Total Duration**: ~70 hours of dexterous hand-object interactions
- **Real Demonstrations**: ~0.7 hours of human mocap data (1/100 of collected data)
- **Expansion Ratio**: 100× from real to simulated data
- **Manipulation Types**: 21 types based on Cutkosky taxonomy
- **Objects**: 30 objects (geometric primitives + YCB objects)
- **Capture Rate**: 100 Hz optical motion capture

## Manipulation Coverage

The dataset spans four primary grasp categories:

- **Power Grasps**: Full-hand wrapping grips
- **Intermediate Grasps**: Mixed precision-power combinations
- **Precision Grasps**: Fingertip-based manipulation
- **In-Hand Manipulation**: Object reorientation and repositioning

All 21 manipulation types follow the Cutkosky grasp taxonomy.

## Data Modalities

Each frame includes:

- **RGB-D Data**: Multi-view color and depth images
- **Hand Pose**: MANO hand parameters with high-precision tracking
- **Object State**: 6-DoF pose and object wrenches
- **Annotations**: Per-frame labels and metadata

**Note**: Contact force data is not included in v0.1. Contact forces will be available in future releases.

## Data Pipeline

The dataset is generated through three stages:

1. **Real Capture**: Optical motion capture of human demonstrations at 30 Hz
2. **Force Reconstruction**: RL-based physics simulation to infer contact forces
3. **Physics Validation**: Verification of contact points, forces, and object dynamics

This hybrid approach provides contact information impossible to observe directly in real-world scenarios while maintaining physical accuracy.

## Installation

```bash
pip install datasets huggingface_hub
```

For image processing and visualization:

```bash
pip install pillow numpy torch
```

Authenticate with HuggingFace (required for private datasets):

```bash
huggingface-cli login
```

Or set your token as an environment variable:

```bash
export HF_TOKEN="your_token_here"
```

## Quick Start

### Data Structure

```json
{
  "trajectory_meta_data": {
    "generated_data": "int",
    "data_fps": "int",
    "mocap_raw_data_source": {
      "operator": "str",
      "object": "str",
      "gesture": "str"
    },
    "total_frames": "int",
    "mano_hand_shape": "(10,)"
    //...
  },
  "sequence_info": {
    "timestamp": "(T,)",
    "hand_joint": {
      "position": "(T, 3)",
      "rotation": "(T, 3)",
      "finger_pose": "(T, 48)"
    },
    "object_info": {
      "pose": "(T, 6)"
    },
    "mano_model_output": {
      "joints": "(T, 63)"
    }
  }
}
```
### Visualization

Visualize trajectories using the **mocap_loader**:

```bash
# Install dependencies
pip install open3d trimesh scipy

# Visualize trajectory
python -m hand_trajectory_loader.examples.visualize_trajectory \
    dataset.parquet 0 \
    --mano-model assets/mano/models/MANO_RIGHT.pkl \
    --object assets/objects/cube1.stl \
    --show-joints
```

Controls: **SPACE** pause/resume, **M** toggle hand mesh, **O** toggle object, **Q** quit

## Version Information

**v0.1 (Test Release)** includes:
- 1% of collected real human demonstration data
- MANO hand parameters
- Object pose data
- Manipulation type annotations

**Coming in future releases**:
- Complete dataset (100× larger than v0.1)
- Contact force data with physics validation
- Additional objects and manipulation types
- Extended annotations and metadata

## Contact

**Research Collaboration**
Academic inquiries: lyw@dex-robot.com

**Business Inquiries**
Business collaboration: info@dex-robot.com

**Website**
https://www.dex-robot.com/en
https://dexcanvas.github.io/

## Citation

```bibtex
@article{dexcanvas2025,
  title={DexCanvas: A Large-Scale Hybrid Dataset for Dexterous Manipulation},
  author={DexRobot Team},
  year={2025},
  eprint={2510.15786},
  archivePrefix={arXiv},
  url={https://arxiv.org/abs/2510.15786}
}
```

## License

This dataset is released under the Open Database License (ODbL).

---

**Developed by DexRobot Team**
Last Updated: October 2025