|
|
--- |
|
|
license: cc-by-nc-4.0 |
|
|
--- |
|
|
# GestureHYDRA: Semantic Co-speech Gesture Synthesis via Hybrid Modality Diffusion Transformer and Cascaded-Synchronized Retrieval-Augmented Generation.ICCV 2025 |
|
|
|
|
|
|
|
|
|
|
|
## Dataset description |
|
|
- The Streamer dataset we proposed consists of 281 anchors and 20,969 clips of data. |
|
|
- The training set contains 19,051 clips of data and 269 anchors. |
|
|
- The seen test set contains 920 clips of data, where the anchor ID has appeared in the training set. |
|
|
- The unseen test set contains 998 clips of data, where the anchor ID has never appeared in the training set. |
|
|
|
|
|
## Dataset structure |
|
|
The dataset contains three folders: *train*, *test_seen*, and *test_unseen*. Let's *train* train as an example for introduction. |
|
|
|
|
|
```plaintext |
|
|
-train/ |
|
|
├── audios/ |
|
|
│ ├── {anchor_id}/ |
|
|
│ │ └── {video_name_md5}/ |
|
|
│ │ └── {start_time}_{end_time}.wav |
|
|
├── gestures/ |
|
|
│ ├── {anchor_id}/ |
|
|
│ │ └── {video_name_md5}/ |
|
|
│ │ └── {start_time}_{end_time}.pkl |
|
|
``` |
|
|
The audio data in the audios folder corresponds one-to-one with the human motion data in the gestures folder. |
|
|
|
|
|
|
|
|
Data contained in the pkl file: |
|
|
- width, height: the video width and height |
|
|
- center: the center point of the video |
|
|
- batch_size: the sequence length |
|
|
- camera_transl: the displacement of the camera |
|
|
- focal_length: the pixel focal length of a camera |
|
|
- body_pose_axis: (bs, 21x3) |
|
|
- jaw_pose: (bs,3) |
|
|
- betas: (1,10) |
|
|
- global_orient: (bs,3) |
|
|
- transl: (bs,3) |
|
|
- left_hand_pose: (bs,15x3) |
|
|
- right_hand_pose: (bs,15x3) |
|
|
- leye_pose: (bs,3) |
|
|
- reye_pose: (bs,3) |
|
|
- pose_embedding: (bs,32) |