Datasets:

License:
HORA / README.md
YueMingLee's picture
docs: fix Robot fields markdown format
5c44755
metadata
license: apache-2.0

HORA: Hand–Object to Robot Action Dataset

Dataset Summary

HORA (Hand–Object to Robot Action) is a large-scale multimodal dataset that converts human hand–object interaction (HOI) demonstrations into robot-usable supervision for cross-embodiment learning. It combines HOI-style annotations (e.g., MANO hand parameters, object pose, contact) with embodied-robot learning signals (e.g., robot observations, end-effector trajectories) under a unified canonical action space.

HORA is constructed from three sources/subsets:

  1. HORA(Mocap): custom multi-view motion capture system with tactile-sensor gloves (includes tactile maps).
  2. HORA(Recordings): custom RGB(D) HOI recording setup (no tactile).
  3. HORA(Public Dataset): derived from multiple public HOI datasets and retargeted to robot embodiments (6/7-DoF arms).

Overall scale: ~150k trajectories across all subsets.


Key Features

  • Unified multimodal representation across subsets, covering both HOI analysis and downstream robotic learning.
  • HOI modalities: MANO hand parameters (pose/shape + global transform), object 6DoF pose, object assets, hand–object contact annotations.
  • Robot modalities: wrist-view & third-person observations, and end-effector pose trajectories for robotic arms, all mapped to a canonical action space.
  • Tactile (mocap subset): dense tactile map for both hand and object (plus object pose & assets).

Dataset Statistics

Subset Tactile #Trajectories Notes
HORA(Mocap) 63,141 6-DoF object pose + assets + tactile map
HORA(Recordings) 23,560 6-DoF object pose + assets
HORA(Public Dataset) 66,924 retargeted cross-embodiment robot modalities
Total ~150k

Supported Tasks and Use Cases

HORA is suitable for:

  • Imitation Learning (IL) / Visuomotor policy learning
  • Vision–Language–Action (VLA) model training and evaluation
  • HOI-centric research: contact analysis, pose/trajectory learning, hand/object dynamics

Data Format

Example Episode Structure

Each episode/trajectory may include:

HOI fields

  • hand_mano: MANO parameters (pose/shape, global rotation/translation)
  • object_pose_6d: 6DoF object pose sequence
  • contact: hand–object contact annotations
  • object_asset: mesh/texture id or path

Robot fields

  • Global Attributes

    • task_description: Natural language instruction for the task (stored as HDF5 attribute).
    • total_demos: Total number of trajectories in the file.
  • Observations (obs group)

    • agentview_rgb: JPEG byte stream (variable length uint8). Decodes to (T, 480, 640, 3).
    • eye_in_hand_{side}_rgb: JPEG byte stream (variable length uint8). Decodes to (T, 480, 640, 3).
    • {prefix}_joint_states: Arm joint positions in radians. Shape (T, N_dof).
    • {prefix}_gripper_states: Gripper joint positions. Shape (T, N_grip).
    • {prefix}_eef_pos: End-effector position in Robot Base Frame. Shape (T, 3).
    • {prefix}_eef_quat: End-effector orientation (w, x, y, z) in Robot Base Frame. Shape (T, 4).
    • object_{name}_pos: Object ground truth position in World Frame. Shape (T, 3).
    • object_{name}_quat: Object ground truth orientation (w, x, y, z) in World Frame. Shape (T, 4).
  • Actions & States

    Note: For multi-robot setups, the fields below concatenate data from all robots in order (e.g., [robot0, robot1]).

    • actions: Joint-space control targets. Shape (T, N_dof + 1). Format: [joint_positions, normalized_gripper] where gripper is in [0, 1].
    • actions_ee: Cartesian control targets. Shape (T, 7). Format: [pos (3), axis-angle (3), normalized_gripper (1)].
    • robot_states: Robot base pose in World Frame. Shape (T, 7 * N_robots). Format: [pos (3), quat (4)] per robot, quat is (w, x, y, z).

Tactile fields (mocap only)

  • tactile_hand: dense tactile map (time × sensors/vertices)
  • tactile_object: dense tactile map