simplexsigil2 commited on
Commit
2c8d061
·
verified ·
1 Parent(s): 8351144

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +194 -3
README.md CHANGED
@@ -1,3 +1,194 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # OmniFall: A Unified Benchmark for Staged-to-Wild Fall Detection
2
+
3
+ [![License: CC BY-NC-SA 4.0](https://img.shields.io/badge/License-CC%20BY--NC--SA%204.0-lightgrey.svg)](https://creativecommons.org/licenses/by-nc-sa/4.0/)
4
+
5
+ This repository contains the annotation and split definitions for OmniFall, a comprehensive benchmark that unifies eight public indoor fall datasets under a consistent ten-class annotation scheme, complemented by the OOPS-Fall benchmark of genuine accidents captured in the wild.
6
+
7
+ **Paper:** [OmniFall: A Unified Staged-to-Wild Benchmark for Human Fall Detection]()
8
+
9
+ ## Overview
10
+
11
+ Falls are the leading cause of fatal injuries among older adults worldwide. While the mechanical event of falling lasts only a fraction of a second, the critical health risk often comes from the ensuing "fallen" state—when a person remains on the ground, potentially injured and unable to call for help.
12
+
13
+ OmniFall addresses three critical limitations in current fall detection research:
14
+
15
+ 1. **Unified Taxonomy:** Rather than binary fall/no-fall classification, we provide a consistent ten-class scheme across datasets that distinguishes transient actions (fall, sit down, lie down, stand up) from their static outcomes (fallen, sitting, lying, standing).
16
+
17
+ 2. **Combined Benchmark:** We unify eight public datasets (14+ hours of video, 112 subjects, 31 camera views) into a single benchmark with standardized train/val/test splits.
18
+
19
+ 3. **In-the-Wild Evaluation:** We include OOPS-Fall, curated from genuine accident videos of the OOPS dataset to test generalization to real-world conditions.
20
+
21
+ ## Datasets
22
+
23
+ This benchmark includes annotations for the following datasets:
24
+
25
+ 1. **CMDFall** (7h 7m) - Multi-subject, multi-view dataset
26
+ 2. **UP Fall** (4h 35m) - Multiple subjects with varied camera views
27
+ 3. **Le2i** (47m) - Multiple room settings
28
+ 4. **GMDCSA24** (21m) - Varied lighting conditions
29
+ 5. **CAUCAFall** (16m) - Includes day/night recordings
30
+ 6. **OCCU** (14m) - Depth camera recordings
31
+ 7. **EDF** (13m) - Depth camera recordings
32
+ 8. **MCFD** (12m) - Single subject, eight camera views
33
+ 9. **OOPS-Fall** - Curated subset of genuine accidents from the OOPS dataset
34
+
35
+ ## Structure
36
+
37
+ The repository is organized as follows:
38
+
39
+ - `labels/` - CSV files containing frame-level annotations for each dataset as well as label2id.csv
40
+ - `splits/` - Train/validation/test splits for cross-subject (CS) and cross-view (CV) evaluation
41
+ - `splits/cs/` - Cross-subject splits, where training, validation, and test sets contain different subjects
42
+ - `splits/cv/` - Cross-view splits, where training, validation, and test sets contain different camera views
43
+ - `dataset.yaml` - Hugging Face datasets configuration with multiple evaluation protocols
44
+
45
+ ### Label Format
46
+
47
+ Each label file in the `labels/` directory follows this format:
48
+
49
+ ```
50
+ path,label,start,end,subject,cam
51
+ path/to/clip,class_id,start_time,end_time,subject_id,camera_id
52
+ ```
53
+
54
+ Where:
55
+ - `path`: Relative path to the video, given the respective dataset root.
56
+ - `label`: Class ID (0-9) corresponding to one of the ten activity classes:
57
+ - 0: walk
58
+ - 1: fall
59
+ - 2: fallen
60
+ - 3: sit_down
61
+ - 4: sitting
62
+ - 5: lie_down
63
+ - 6: lying
64
+ - 7: stand_up
65
+ - 8: standing
66
+ - 9: other
67
+ - `start`: Start time of the segment (in seconds)
68
+ - `end`: End time of the segment (in seconds)
69
+ - `subject`: Subject ID
70
+ - `cam`: Camera view ID
71
+
72
+ For OOPS-Fall, only fall segments and non-fall segments are labeled; non-falls are labels as "other", independent of the underlying content, as long as it is not a fall.
73
+
74
+ ### Split Format
75
+
76
+ Split files in the `splits/` directory list the video segments included in each partition:
77
+
78
+ ```
79
+ path
80
+ path/to/clip
81
+ ```
82
+
83
+ ## Evaluation Protocols
84
+
85
+ We provide multiple evaluation configurations via the `dataset.yaml` file:
86
+
87
+ ### Basic Configurations
88
+ - `default`: Access to all dataset labels
89
+ - `cs`: Cross-subject splits for all datasets
90
+ - `cv`: Cross-view splits for all datasets
91
+
92
+ ### Individual Dataset Configurations
93
+ - `caucafall`, `cmdfall`, `edf`, `gmdcsa24`, `le2i`, `mcfd`, `occu`, `up_fall`, `OOPS`:
94
+ Access to individual datasets with their respective cross-subject splits
95
+
96
+ ### Multi-Dataset Evaluation Protocols
97
+ - `cs-staged`: Cross-subject splits combined across all staged datasets
98
+ - `cv-staged`: Cross-view splits combined across all staged datasets
99
+ - `cs-staged-wild`: Train and validate on staged datasets with cross-subject splits, test on OOPS-Fall
100
+ - `cv-staged-wild`: Train and validate on staged datasets with cross-view splits, test on OOPS-Fall
101
+
102
+ ## Usage
103
+
104
+ To use this dataset with the Hugging Face datasets library:
105
+
106
+ ```python
107
+ from datasets import load_dataset
108
+
109
+ # Load the entire dataset with default configuration
110
+ dataset = load_dataset("omnifall")
111
+
112
+ # Use cross-subject (CS) evaluation protocol
113
+ cs_dataset = load_dataset("omnifall", "cs")
114
+ print(f"Train: {len(cs_dataset['train'])} samples")
115
+ print(f"Validation: {len(cs_dataset['validation'])} samples")
116
+ print(f"Test: {len(cs_dataset['test'])} samples")
117
+
118
+ # Use cross-view (CV) evaluation protocol
119
+ cv_dataset = load_dataset("omnifall", "cv")
120
+
121
+ # Use staged-to-wild evaluation protocol (train on staged datasets, test on OOPS)
122
+ staged_to_wild = load_dataset("omnifall", "cs-staged-wild")
123
+
124
+ # Use individual dataset
125
+ cmdfall = load_dataset("omnifall", "cmdfall")
126
+
127
+ # Access specific fields from the dataset
128
+ for item in dataset["train"][:5]:
129
+ print(f"Path: {item['path']}, Label: {item['label']}")
130
+ ```
131
+
132
+ ## Experiment Examples
133
+
134
+ ### Cross-Subject Fall Detection
135
+
136
+ ```python
137
+ from datasets import load_dataset
138
+ import torch
139
+ from torch.utils.data import DataLoader
140
+
141
+ # Load the cross-subject evaluation protocol
142
+ dataset = load_dataset("omnifall", "cs-staged")
143
+
144
+ # Preprocess and create dataloaders
145
+ def preprocess(examples):
146
+ # Your preprocessing code here
147
+ return examples
148
+
149
+ processed_dataset = dataset.map(preprocess, batched=True)
150
+ train_dataloader = DataLoader(processed_dataset["train"], batch_size=32, shuffle=True)
151
+ val_dataloader = DataLoader(processed_dataset["validation"], batch_size=32)
152
+ test_dataloader = DataLoader(processed_dataset["test"], batch_size=32)
153
+
154
+ # Train and evaluate your model
155
+ ```
156
+
157
+ ### Staged-to-Wild Generalization
158
+
159
+ ```python
160
+ from datasets import load_dataset
161
+
162
+ # Load the staged-to-wild evaluation protocol
163
+ dataset = load_dataset("omnifall", "cs-staged-wild")
164
+
165
+ # Train on staged data
166
+ train_data = dataset["train"]
167
+ val_data = dataset["validation"]
168
+
169
+ # Evaluate on wild data
170
+ wild_test_data = dataset["test"]
171
+ ```
172
+
173
+ ## Citation
174
+
175
+ If you use OmniFall in your research, please cite our paper:
176
+
177
+ ```bibtex
178
+ @inproceedings{omnifall2025,
179
+ title={OmniFall: A Unified Staged-to-Wild Benchmark for Human Fall Detection},
180
+ author={},
181
+ booktitle={},
182
+ year={2025},
183
+ }
184
+ ```
185
+
186
+ ## License
187
+
188
+ The annotations and split definitions in this repository are released under [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-nc-sa/4.0/).
189
+
190
+ The original video data belongs to their respective owners and should be obtained from the original sources.
191
+
192
+ ## Contact
193
+
194
+ For questions about the dataset, please contact [david.schneider@kit.edu].