Datasets:

License:
YuhongZhang commited on
Commit
7721d4c
·
verified ·
1 Parent(s): 2b0165f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +79 -3
README.md CHANGED
@@ -1,3 +1,79 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ # HORA: Hand–Object to Robot Action Dataset
6
+
7
+ ## Dataset Summary
8
+
9
+ **HORA (Hand–Object to Robot Action)** is a large-scale **multimodal** dataset that converts human **hand–object interaction (HOI)** demonstrations into **robot-usable supervision** for cross-embodiment learning. It combines HOI-style annotations (e.g., MANO hand parameters, object pose, contact) with embodied-robot learning signals (e.g., robot observations, end-effector trajectories) under a unified canonical action space. :contentReference[oaicite:5]{index=5} :contentReference[oaicite:6]{index=6}
10
+
11
+ HORA is constructed from **three sources/subsets**:
12
+
13
+ 1. **HORA(Mocap)**: custom multi-view motion capture system with **tactile-sensor gloves** (includes tactile maps).
14
+ 2. **HORA(Recordings)**: custom RGB(D) HOI recording setup (no tactile).
15
+ 3. **HORA(Public Dataset)**: derived from multiple public HOI datasets and retargeted to robot embodiments (6/7-DoF arms).
16
+
17
+ Overall scale: **~150k trajectories** across all subsets.
18
+
19
+ ---
20
+
21
+ ## Key Features
22
+
23
+ - **Unified multimodal representation** across subsets, covering both HOI analysis and downstream robotic learning.
24
+ - **HOI modalities**: MANO hand parameters (pose/shape + global transform), object 6DoF pose, object assets, hand–object contact annotations.
25
+ - **Robot modalities**: wrist-view & third-person observations, and end-effector pose trajectories for robotic arms, all mapped to a canonical action space.
26
+ - **Tactile** (mocap subset): dense tactile map for both hand and object (plus object pose & assets).
27
+
28
+ ---
29
+
30
+ ## Dataset Statistics
31
+
32
+ | Subset | Tactile | #Trajectories | Notes |
33
+ | --- | --- | --- | --- |
34
+ | HORA(Mocap) | ✅ | 63,141 | 6-DoF object pose + assets + tactile map |
35
+ | HORA(Recordings) | ❌ | 23,560 | 6-DoF object pose + assets |
36
+ | HORA(Public Dataset) | ❌ | 66,924 | retargeted cross-embodiment robot modalities |
37
+ | **Total** | | **~150k** | |
38
+
39
+ ## Supported Tasks and Use Cases
40
+
41
+ HORA is suitable for:
42
+
43
+ - **Imitation Learning (IL)** / **Visuomotor policy learning**
44
+ - **Vision–Language–Action (VLA)** model training and evaluation
45
+ - HOI-centric research: contact analysis, pose/trajectory learning, hand/object dynamics
46
+
47
+
48
+ ---
49
+
50
+ ## Data Format
51
+
52
+
53
+
54
+ ### Example Episode Structure
55
+
56
+ Each episode/trajectory may include:
57
+
58
+ **HOI fields**
59
+
60
+ - `hand_mano`: MANO parameters (pose/shape, global rotation/translation)
61
+ - `object_pose_6d`: 6DoF object pose sequence
62
+ - `contact`: hand–object contact annotations
63
+ - `object_asset`: mesh/texture id or path
64
+
65
+ **Robot fields**
66
+
67
+ - `obs_wrist_rgb`
68
+ - `obs_third_rgb`
69
+ - `ee_pose`: end-effector pose trajectory (SE(3))
70
+ - `gripper`: gripper open/close command (optional)
71
+ - `action_space`: canonical action space metadata
72
+
73
+ **Tactile fields (mocap only)**
74
+
75
+ - `tactile_hand`: dense tactile map (time × sensors/vertices)
76
+ - `tactile_object`: dense tactile map
77
+
78
+
79
+ ---