File size: 5,945 Bytes
f9a9ce1 06c13a9 f9a9ce1 06c13a9 f9a9ce1 b063b13 f9a9ce1 b063b13 06c13a9 f9a9ce1 f38109c f9a9ce1 b063b13 f8c51c6 f38109c d4ee670 1579185 b86c2c7 1579185 b063b13 5db77d1 b063b13 a5dfafa f38109c b063b13 061ef3c | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 | ---
dataset_info:
features:
- name: uid
dtype: string
- name: video_id
dtype: string
- name: start_second
dtype: float32
- name: end_second
dtype: float32
- name: caption
dtype: string
- name: fx
dtype: float32
- name: fy
dtype: float32
- name: cx
dtype: float32
- name: cy
dtype: float32
- name: vid_w
dtype: int32
- name: vid_h
dtype: int32
- name: annotation
list:
- name: mano_params
struct:
- name: global_orient
list: float32
- name: hand_pose
list: float32
- name: betas
list: float32
- name: is_right
dtype: bool
- name: keypoints_3d
list: float32
- name: keypoints_2d
list: float32
- name: vertices
list: float32
- name: box_center
list: float32
- name: box_size
dtype: float32
- name: camera_t
list: float32
- name: focal_length
list: float32
splits:
- name: train
num_examples: 241912
- name: test
num_examples: 5108
configs:
- config_name: default
data_files:
- split: train
path: EgoHaFL_train.csv
- split: test
path: EgoHaFL_test.csv
license: mit
language:
- en
pretty_name: EgoHaFL:Egocentric 3D Hand Forecasting Dataset with Language Instruction
size_categories:
- 200K<n<300K
tags:
- embodied-ai
- robotic
- egocentric
- 3d-hand
- forecasting
- hand-pose
---
# **EgoHaFL: Egocentric 3D Hand Forecasting Dataset with Language Instruction**
**EgoHaFL** is a dataset designed for **egocentric (first-person) 3D hand forecasting** with accompanying **natural language instructions**.
It contains short video clips, text descriptions, camera intrinsics, and detailed MANO-based 3D hand annotations.
The dataset supports research in **3D hand forecasting**, **hand pose estimation**, **hand–object interaction understanding**, and **video–language modeling**.

[](https://arxiv.org/pdf/2511.18127)
[](https://huggingface.co/ut-vision/SFHand)
[](https://github.com/ut-vision/SFHand)
---
## 📦 **Dataset Contents**
### **1. Metadata CSV Files**
* `EgoHaFL_train.csv`
* `EgoHaFL_test.csv`
Each row corresponds to one sample and contains:
| Field | Description |
| ---------------- | ------------------------------------------ |
| `uid` | Unique sample identifier |
| `video_id` | Source video identifier |
| `start_second` | Start time of the clip (seconds) |
| `end_second` | End time of the clip (seconds) |
| `caption` | Natural language instruction / description |
| `fx`, `fy` | Camera focal lengths |
| `cx`, `cy` | Principal point |
| `vid_w`, `vid_h` | Original video resolution |
---
### **2. 3D Hand Annotations (EgoHaFL_lmdb)**
The folder `EgoHaFL_lmdb` stores all 3D annotations in **LMDB format**.
* **Key**: `uid`
* **Value**: a **list of length 16**, representing uniformly sampled frames across a **3-second video segment**.
Each of the 16 elements is a dictionary containing:
* `mano_params`
* `global_orient (n, 1, 3 ,3)`
* `hand_pose (n, 15, 3, 3)`
* `betas (n, 10)`
* `is_right (n,)`
* `keypoints_3d (n, 21, 3)`
* `keypoints_2d (n, 21, 2)`
* `vertices (n, 778, 3)`
* `box_center (n, 2)`
* `box_size (n,)`
* `camera_t (n, 3)` *3D hand position in camera coordinate*
* `focal_length (n, 2)`
Here, `n` denotes the number of hands present in each frame, which may vary across frames. When no hands are detected, the dictionary is empty.
---
## 🌳 **Annotation Structure (Tree View)**
Below is the hierarchical structure for a single annotation entry (`uid → 16-frame list → per-frame dict`):
```
<uid>
└── list (length = 16)
├── [0]
│ ├── mano_params
│ │ ├── global_orient
│ │ ├── hand_pose
│ │ └── betas
│ ├── is_right
│ ├── keypoints_3d
│ ├── keypoints_2d
│ ├── vertices
│ ├── box_center
│ ├── box_size
│ ├── camera_t
│ └── focal_length
├── [1]
│ └── ...
├── [2]
│ └── ...
└── ...
```
---
## 🎥 **Source of Video Data**
The video clips used in **EgoHaFL** originate from the **Ego4D V1** dataset.
For our experiments, we use the **original-length videos compressed to 224p resolution** to ensure efficient storage and training.
Official Ego4D website:
🔗 **[https://ego4d-data.org/](https://ego4d-data.org/)**
---
## 🧩 **Example of Use**
For details on how to load and use the EgoHaFL dataset,
please refer to the **dataloader implementation** in our GitHub repository:
🔗 **[https://github.com/ut-vision/SFHand](https://github.com/ut-vision/SFHand)**
---
## 🧠 **Supported Research Tasks**
* Egocentric 3D hand forecasting
* Hand motion prediction and trajectory modeling
* 3D hand pose estimation
* Hand–object interaction understanding
* Video–language multimodal modeling
* Temporal reasoning with 3D human hands
---
## 📚 Citation
If you use this dataset or find it helpful in your research, please cite:
```latex
@article{liu2025sfhand,
title={SFHand: A Streaming Framework for Language-guided 3D Hand Forecasting and Embodied Manipulation},
author={Liu, Ruicong and Huang, Yifei and Ouyang, Liangyang and Kang, Caixin and and Sato, Yoichi},
journal={arXiv preprint arXiv:2511.18127},
year={2025}
}
```
|