Datasets:

License:
File size: 4,158 Bytes
7721d4c
 
 
 
 
 
 
 
45e6c99
7721d4c
 
 
 
 
45e6c99
7721d4c
45e6c99
7721d4c
 
 
 
 
45e6c99
 
 
 
7721d4c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45e6c99
 
 
 
 
5c44755
 
 
 
 
 
 
 
45e6c99
 
5c44755
 
 
 
45e6c99
5c44755
 
 
7721d4c
 
 
 
 
 
45e6c99
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
---
license: apache-2.0
---

# HORA: Hand–Object to Robot Action Dataset

## Dataset Summary

**HORA (Hand–Object to Robot Action)** is a large-scale **multimodal** dataset that converts human **hand–object interaction (HOI)** demonstrations into **robot-usable supervision** for cross-embodiment learning. It combines HOI-style annotations (e.g., MANO hand parameters, object pose, contact) with embodied-robot learning signals (e.g., robot observations, end-effector trajectories) under a unified canonical action space.

HORA is constructed from **three sources/subsets**:

1. **HORA(Mocap)**: custom multi-view motion capture system with **tactile-sensor gloves** (includes tactile maps).
2. **HORA(Recordings)**: custom RGB(D) HOI recording setup (no tactile).
3. **HORA(Public Dataset)**: derived from multiple public HOI datasets and retargeted to robot embodiments (6/7-DoF arms).

Overall scale: **~150k trajectories** across all subsets.

---

## Key Features

- **Unified multimodal representation** across subsets, covering both HOI analysis and downstream robotic learning.
- **HOI modalities**: MANO hand parameters (pose/shape + global transform), object 6DoF pose, object assets, hand–object contact annotations.
- **Robot modalities**: wrist-view & third-person observations, and end-effector pose trajectories for robotic arms, all mapped to a canonical action space.
- **Tactile** (mocap subset): dense tactile map for both hand and object (plus object pose & assets).

---

## Dataset Statistics

| Subset | Tactile | #Trajectories | Notes |
| --- | --- | --- | --- |
| HORA(Mocap) | ✅ | 63,141 | 6-DoF object pose + assets + tactile map |
| HORA(Recordings) | ❌ | 23,560 | 6-DoF object pose + assets |
| HORA(Public Dataset) | ❌ | 66,924 | retargeted cross-embodiment robot modalities |
| **Total** |  | **~150k** |  |

## Supported Tasks and Use Cases

HORA is suitable for:

- **Imitation Learning (IL)** / **Visuomotor policy learning**
- **Vision–Language–Action (VLA)** model training and evaluation
- HOI-centric research: contact analysis, pose/trajectory learning, hand/object dynamics


---

## Data Format



### Example Episode Structure

Each episode/trajectory may include:

**HOI fields**

- `hand_mano`: MANO parameters (pose/shape, global rotation/translation)
- `object_pose_6d`: 6DoF object pose sequence
- `contact`: hand–object contact annotations
- `object_asset`: mesh/texture id or path

**Robot fields**

- **Global Attributes**
  - `task_description`: Natural language instruction for the task (stored as HDF5 attribute).
  - `total_demos`: Total number of trajectories in the file.

- **Observations** (`obs` group)
  - `agentview_rgb`: JPEG byte stream (variable length `uint8`). Decodes to `(T, 480, 640, 3)`.
  - `eye_in_hand_{side}_rgb`: JPEG byte stream (variable length `uint8`). Decodes to `(T, 480, 640, 3)`.
  - `{prefix}_joint_states`: Arm joint positions in radians. Shape `(T, N_dof)`.
  - `{prefix}_gripper_states`: Gripper joint positions. Shape `(T, N_grip)`.
  - `{prefix}_eef_pos`: End-effector position in **Robot Base Frame**. Shape `(T, 3)`.
  - `{prefix}_eef_quat`: End-effector orientation `(w, x, y, z)` in **Robot Base Frame**. Shape `(T, 4)`.
  - `object_{name}_pos`: Object ground truth position in **World Frame**. Shape `(T, 3)`.
  - `object_{name}_quat`: Object ground truth orientation `(w, x, y, z)` in **World Frame**. Shape `(T, 4)`.

- **Actions & States**
  > **Note:** For multi-robot setups, the fields below concatenate data from all robots in order (e.g., `[robot0, robot1]`).
  - `actions`: Joint-space control targets. Shape `(T, N_dof + 1)`.
    *Format: `[joint_positions, normalized_gripper]` where gripper is in `[0, 1]`.*
  - `actions_ee`: Cartesian control targets. Shape `(T, 7)`.
    *Format: `[pos (3), axis-angle (3), normalized_gripper (1)]`.*
  - `robot_states`: Robot base pose in **World Frame**. Shape `(T, 7 * N_robots)`.
    *Format: `[pos (3), quat (4)]` per robot, quat is `(w, x, y, z)`.*

**Tactile fields (mocap only)**

- `tactile_hand`: dense tactile map (time × sensors/vertices)
- `tactile_object`: dense tactile map


---