PSewmuthu commited on
Commit
0f35820
Β·
1 Parent(s): b9cef59

Add README file

Browse files
Files changed (1) hide show
  1. README.md +120 -1
README.md CHANGED
@@ -1,3 +1,122 @@
1
  ---
2
- license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ pretty_name: How2Sign Holistic
3
+ language: en
4
+ license:
5
+ - mit
6
+ tags:
7
+ - sign-language
8
+ - asl
9
+ - mediapipe
10
+ - holistic
11
+ - pose-landmarks
12
+ - hand-landmarks
13
+ - face-landmarks
14
+ - gesture-recognition
15
+ - sequence-modeling
16
+ - time-series
17
+ - computer-vision
18
+ - deep-learning
19
+ source_datasets:
20
+ - Duarte_CVPR2021/How2Sign
21
+ task_categories:
22
+ - feature-extraction
23
+ - translation
24
+ task_ids:
25
+ - pose-estimation
26
+ - conversational
27
+ citation:
28
+ - "@inproceedings{Duarte_CVPR2021,
29
+ title={{How2Sign: A Large-scale Multimodal Dataset for Continuous American Sign Language}},
30
+ author={Duarte, Amanda and Palaskar, Shruti and Ventura, Lucas and Ghadiyaram, Deepti and DeHaan, Kenneth and
31
+ Metze, Florian and Torres, Jordi and Giro-i-Nieto, Xavier},
32
+ booktitle={Conference on Computer Vision and Pattern Recognition (CVPR)},
33
+ year={2021}
34
+ }"
35
+ - "@misc{MediaPipe,
36
+ title={MediaPipe},
37
+ author={Google Inc.},
38
+ year={2020},
39
+ url={https://mediapipe.dev/}
40
+ }"
41
  ---
42
+
43
+ # How2Sign Holistic
44
+
45
+ ### Mediapipe Holistic Landmark Features Extracted from the How2Sign ASL Dataset
46
+
47
+ ## Overview
48
+
49
+ **How2Sign Holistic** is a curated dataset providing frame-level Mediapipe Holistic landmarks extracted from the full How2Sign American Sign Language corpus. Each sentence-level video clip has pose, face, and hand landmark sequences stored as `.npy` files.
50
+
51
+ This dataset is designed to support research in:
52
+
53
+ - ASL recognition and translation
54
+ - Pose-based sign generation
55
+ - Sequence and time-series modeling
56
+ - Gesture understanding
57
+ - Multiview motion analysis
58
+
59
+ ## Base Directory
60
+
61
+ **`how2sign_holistic_features/`** is the root folder containing all splits and metadata.
62
+
63
+ ## Sources
64
+
65
+ The original data comes from the **How2Sign dataset** (Duarte et al., CVPR 2021), a large-scale multimodal American Sign Language dataset sourced from YouTube videos.
66
+
67
+ ## Collection Methodology
68
+
69
+ - Sentence-level clips were extracted from the original videos according to How2Sign protocol.
70
+ - Frame-level landmarks were extracted using **Google Mediapipe Holistic** (pose, face, hands).
71
+ - Each clip saved as `.npy` with frontal and side views.
72
+ - Metadata CSVs map clips to sentences, start/end timestamps, and video identifiers.
73
+ - CSVs can be opened in pandas: `pd.read_csv('filename.csv', sep='\t')`
74
+
75
+ ## Dataset Structure
76
+
77
+ ```
78
+ how2sign_holistic_features/
79
+ β”‚
80
+ β”œβ”€β”€ metadata/ # Original How2Sign metadata (CSV files)
81
+ β”‚ β”œβ”€β”€ how2sign_realigned_train.csv
82
+ β”‚ β”œβ”€β”€ how2sign_realigned_val.csv
83
+ β”‚ β”œβ”€β”€ how2sign_realigned_test.csv
84
+ β”‚ β”œβ”€β”€ how2sign_train.csv
85
+ β”‚ β”œβ”€β”€ how2sign_val.csv
86
+ β”‚ └── how2sign_test.csv
87
+ β”‚
88
+ β”œβ”€β”€ train/ # Training split .npy files
89
+ β”‚ β”œβ”€β”€ frontal/
90
+ β”‚ β”‚ β”œβ”€β”€ <VIDEO_ID>_front_holistic.npy
91
+ β”‚ β”‚ └── ...
92
+ β”‚ └── side/
93
+ β”‚ β”œβ”€β”€ <VIDEO_ID>_side_holistic.npy
94
+ β”‚ └── ...
95
+ β”‚
96
+ β”œβ”€β”€ val/ # Validation split
97
+ β”‚ β”œβ”€β”€ frontal/
98
+ β”‚ └── side/
99
+ β”‚
100
+ └── test/ # Test split
101
+ β”œβ”€β”€ frontal/
102
+ └── side/
103
+ ```
104
+
105
+ ### Notes
106
+
107
+ - `.npy` files contain **frame-level Mediapipe Holistic landmarks**.
108
+ - Frontal and side views are synchronized.
109
+ - Filenames follow: `VIDEO_NAME_START-END-rgb_front/side_holistic.npy`
110
+ - Metadata CSVs map clips to video ID, sentence, start/end timestamps, and How2Sign identifiers.
111
+
112
+ ## Citation
113
+
114
+ If you use this dataset, please cite:
115
+
116
+ Duarte, A., Palaskar, S., Ventura, L., Ghadiyaram, D., DeHaan, K., Metze, F., Torres, J., & Giro-i-Nieto, X.
117
+ **β€œHow2Sign: A Large-scale Multimodal Dataset for Continuous American Sign Language.”**
118
+ _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021._
119
+
120
+ ## Recommended Tags
121
+
122
+ `ASL`, `Sign Language`, `Mediapipe`, `Holistic`, `Pose Landmarks`, `Hand Landmarks`, `Face Landmarks`, `Keypoints`, `Motion Capture`, `Time Series`, `Gesture Recognition`, `Computer Vision`, `Deep Learning`, `Sequence Modeling`