File size: 12,058 Bytes
7ab8c60 9c6ddbf 7ab8c60 7abcc09 7ab8c60 9c6ddbf 7abcc09 9c6ddbf 561f292 9c6ddbf 7abcc09 9c6ddbf 7abcc09 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 871293f 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 7abcc09 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 40cd64b 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 40cd64b 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 9c6ddbf 7ab8c60 72d2534 7abcc09 72d2534 7ab8c60 9c6ddbf | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 | ---
license: cc-by-nc-4.0
language:
- ase
- bfi
- gsg
- sgd
- fsl
- lsf
- lse
- lis
- lgp
- ngt
- asf
- jsl
- kvk
- csl
- aed
- tsm
- pjm
- rsl
- swl
- dsl
- fse
- nsl
- lsc
- lsm
- bzs
task_categories:
- other
tags:
- sign-language
- pose-estimation
- dwpose
- multilingual
- keypoint
- video-understanding
- sign-language-generation
- sign-language-recognition
- pose-native
size_categories:
- 1M<n<10M
pretty_name: "SignVerse-2M: A Two-Million-Clip Pose-Native Universe of 55+ Sign Languages"
---
<div align="center" style="position: relative; width: 100%; max-width: 1400px; margin: 0 auto 24px auto;">
<img
src="https://signerx.github.io/SignVerse-2M/static/images/background_gallery_big.gif"
alt="SignVerse-2M cover"
style="width: 100%; display: block; border-radius: 18px;"
/>
<div
style="
position: absolute;
inset: 0;
display: flex;
align-items: center;
justify-content: center;
background: linear-gradient(to bottom, rgba(0,0,0,0.18), rgba(0,0,0,0.28));
border-radius: 18px;
"
>
<div
style="
color: white;
font-size: 3.4rem;
font-weight: 800;
letter-spacing: 0.04em;
text-shadow: 0 4px 20px rgba(0,0,0,0.45);
"
>
SignVerse-2M
</div>
</div>
</div>
## SignVerse-2M: A Two-Million-Clip Pose-Native Universe of 55+ Sign Languages
Links: [[Paper]](https://arxiv.org/abs/2605.01720) | [[Data Files]](https://huggingface.co/datasets/SignerX/SignVerse-2M/tree/main/dataset) | [[Project Page]](https://signerx.github.io/SignVerse-2M)
**SignVerse-2M** is a large-scale multilingual pose-native dataset for sign language research. The dataset reorganizes publicly available sign language videos into a unified DWPose-based representation and releases the result as approximately **2 million clips** from **39,196 videos** covering **55+ sign languages**. Rather than distributing raw RGB video, SignVerse-2M provides per-frame body, hand, and face keypoints together with structured subtitle supervision, making the corpus directly usable for pose-conditioned sign language generation, recognition, and translation research.
## Overview
Existing large-scale sign language resources are typically organized as video-text corpora. That format is appropriate for RGB-based recognition or translation, but it is not the most natural interface for modern pose-driven generation pipelines, which increasingly operate on standardized human keypoint controls such as DWPose. SignVerse-2M addresses this mismatch by converting multilingual public sign language videos into a common pose space.
The release is intended to support research questions such as:
- multilingual sign language generation in pose space
- pose-based sign language recognition and translation
- cross-lingual transfer across heterogeneous sign language sources
- benchmarking of sign language motion representations under open-world conditions
## Key Characteristics
| Property | Value |
| --- | --- |
| Dataset name | SignVerse-2M |
| Core representation | DWPose keypoint sequences |
| Videos | 39,196 |
| Clips / subtitle segments | Approximately 2 million |
| Sign languages | 55+ |
| Frame rate | 24 FPS |
| Per-frame keypoints | 18 body + 21 left hand + 21 right hand + 68 face = 128 |
| Source type | Public multilingual sign language videos |
| Raw RGB frames released | No |
| Released supervision | Structured subtitle text and document-level text |
## Why A Pose-Native Release
SignVerse-2M should not be understood as merely a larger multilingual video-text corpus. Its main contribution is the release of a unified pose-native interface for sign language research.
Compared with raw-video releases, the pose-native representation offers three practical advantages:
1. It reduces nuisance variation from background, clothing, and appearance, allowing models to focus more directly on motion.
2. It aligns naturally with contemporary pose-conditioned generation pipelines that already consume DWPose-like controls.
3. It provides a common representation for multilingual benchmarking, making comparisons across methods more interpretable.
## Data Source And Processing
The corpus is built from publicly available multilingual sign language videos ([YouTube-SL-25](https://arxiv.org/abs/2407.11144)), including resources inherited from large public sign language collections such as YouTube-SL-55 and related open web sources. Each video is processed through a unified pipeline that:
1. retrieves metadata and available subtitles,
2. structures subtitle tracks into segment-level and document-level text,
3. decodes the video at 24 FPS,
4. applies DWPose to extract body, hand, and face keypoints frame by frame,
5. packages the outputs into per-video artifacts for public release.
No manual keypoint annotation is provided. The keypoints and subtitles are produced automatically through the preprocessing pipeline.
## Languages
The corpus covers more than 55 sign languages. Major language codes in the current release include:
| Code | Language | Code | Language |
| --- | --- | --- | --- |
| `ase` | American Sign Language | `lsf` | French Sign Language |
| `bfi` | British Sign Language | `lse` | Spanish Sign Language |
| `gsg` | German Sign Language | `lis` | Italian Sign Language |
| `sgd` | Swiss German Sign Language | `lgp` | Portuguese Sign Language |
| `asf` | Australian Sign Language | `ngt` | Sign Language of the Netherlands |
| `jsl` | Japanese Sign Language | `kvk` | Korean Sign Language |
| `csl` | Chinese Sign Language | `bzs` | Brazilian Sign Language |
| `lsm` | Mexican Sign Language | `pjm` | Polish Sign Language |
The language distribution is long-tailed rather than balanced. High-resource languages account for a disproportionate share of the total data volume.
## Repository Structure
The public release is organized around `.tar` shards stored under `dataset/`. Each shard contains per-video directories:
```text
dataset/
Sign_DWPose_NPZ_000001.tar
Sign_DWPose_NPZ_000002.tar
...
```
Within each shard:
```text
{video_id}/
poses.npz
caption.json
{video_id}.complete
```
The main files are:
- `poses.npz`: per-video DWPose payload with frame-wise keypoints
- `caption.json`: structured subtitle and supervision metadata
- `.complete`: completion marker produced by the processing pipeline
## Data Schema
### `poses.npz`
Each `poses.npz` file stores a person-centric per-frame representation. A simplified schema is shown below:
```python
{
"video_id": str,
"fps": float,
"num_frames": int,
"frame_ids": int[T],
"width": int,
"height": int,
"frames": [
{
"num_people": int,
"frame_id": int,
"width": int,
"height": int,
"person_0": {
"body": float[18, 3],
"face": float[68, 3],
"left_hand": float[21, 3],
"right_hand": float[21, 3],
},
# optional additional people:
# "person_1": { ... }
},
...
]
}
```
Keypoint coordinates are stored in pixel space as `(x, y, score)`, where confidence scores lie in `[0, 1]`.
### `caption.json`
```json
{
"video_id": "...",
"sign_language": "ase",
"title": "...",
"duration_s": 312.4,
"segments": [
{ "start": 0.0, "end": 4.2, "text": "..." }
],
"document_text": "...",
"english_source": "native"
}
```
The field `english_source` records whether the English supervision is native or automatically selected from an available translated subtitle track.
## Loading Example
```python
import json
import tarfile
import numpy as np
with tarfile.open("dataset/Sign_DWPose_NPZ_000001.tar") as tar:
tar.extractall("./tmp_signverse")
npz = np.load("./tmp_signverse/{video_id}/poses.npz", allow_pickle=True)
frames = npz["frames"].tolist()
body = frames[0]["person_0"]["body"]
with open("./tmp_signverse/{video_id}/caption.json", "r", encoding="utf-8") as f:
caption = json.load(f)
print(body.shape)
print(caption["segments"][0]["text"])
```
## Visualization And Reproduction
The repository includes scripts for inspecting the released pose files and for reproducing the processing pipeline.
### Visualize one pose file
```bash
python scripts/visualize_dwpose_npz.py \
--npz extracted/{video_id}/poses.npz \
--style openpose \
--out viz/
```
### Reproduce the pipeline
```bash
# Single machine
bash reproduce_independently.sh
# SLURM cluster
bash reproduce_independently_slurm.sh
```
The pipeline is organized into acquisition, subtitle structuring, pose extraction, and upload/publication stages.
## Benchmark Setting
The accompanying paper introduces a multilingual **text-to-pose** benchmark for sign language generation. A generated DWPose sequence is evaluated through back-translation into spoken text, and standard text metrics such as BLEU and ROUGE are reported against the source input. The benchmark repository also provides a **SignDW Transformer** baseline in both small and large model configurations.
For model code and experimental setup, refer to the benchmark repository:
- [https://github.com/SignerX/SignVerse-2M](https://github.com/SignerX/SignVerse-2M)
## Intended Use
The release is intended for research use, including:
- sign language generation from text via pose space
- pose-based sign language translation and recognition
- cross-lingual transfer, adaptation, and benchmarking
- comparison of pose-native motion representations under open-world distributions
The release is not intended for:
- safety-critical interpretation in medical, legal, or emergency settings
- re-identification of individual signers
- claims of full linguistic coverage for any specific sign language
<!--
## Limitations
Users should account for the following limitations when interpreting results:
- Keypoints are extracted automatically and may be noisy under fast motion, occlusion, or multi-person scenes.
- Fine-grained handshape distinctions are only partially captured by the released 21-keypoint hand representation.
- Non-manual linguistic signals such as facial expression and mouthing are only partially represented by 68 face landmarks.
- Subtitle timing and translations are automatically processed and may contain alignment or semantic errors.
- The corpus is language-imbalanced and inherits the long-tail distribution of public web video sources.
- `person_0` is treated as the primary signer, which may be imperfect in multi-signer videos.
-->
## Responsible Use
SignVerse-2M is derived from publicly posted sign language videos. This repository does **not** redistribute raw RGB videos; it releases pose keypoints and structured subtitle text only. Even so, pose sequences may still carry information that can contribute to signer identification when combined with external metadata. Users should treat the corpus as human-subject-derived data and use it responsibly.
The data distribution is also shaped by what is publicly available online. Educational or interpreter-style content may be overrepresented, while conversational, regional, or community-specific signing practices may be underrepresented.
## Citation
If you use SignVerse-2M in academic work, please cite:
```bibtex
@misc{fang2026signverse2mtwomillionclipposenativeuniverse,
title={SignVerse-2M: A Two-Million-Clip Pose-Native Universe of 55+ Sign Languages},
author={Sen Fang and Hongbin Zhong and Yanxin Zhang and Dimitris N. Metaxas},
year={2026},
eprint={2605.01720},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2605.01720},
}
```
## License
The released dataset annotations, pose keypoints, and accompanying metadata are distributed under **CC BY-NC 4.0**.
Source videos are **not** redistributed in this repository and remain subject to the original platform terms and the rights of their respective creators.
|