hrinnnn commited on
Commit
3d347de
·
verified ·
1 Parent(s): 3192c46

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +203 -3
README.md CHANGED
@@ -1,3 +1,203 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pretty_name: PerceptionComp
3
+ license: other
4
+ task_categories:
5
+ - video-question-answering
6
+ - multiple-choice
7
+ language:
8
+ - en
9
+ tags:
10
+ - video
11
+ - benchmark
12
+ - multimodal
13
+ - reasoning
14
+ - video-understanding
15
+ - evaluation
16
+ size_categories:
17
+ - 1K<n<10K
18
+ ---
19
+
20
+ # PerceptionComp
21
+
22
+ PerceptionComp is a benchmark for complex perception-centric video reasoning. It focuses on questions that cannot be solved from a single frame, a short clip, or a shallow caption. Models must revisit visually complex videos, gather evidence across temporally separated segments, and combine multiple perceptual cues before answering.
23
+
24
+ ## Dataset Details
25
+
26
+ ### Dataset Description
27
+
28
+ PerceptionComp contains 1,114 manually annotated five-choice questions associated with 273 videos. The benchmark covers seven categories: outdoor tour, shopping, sport, variety show, home tour, game, and movie.
29
+
30
+ This Hugging Face dataset repository hosts the benchmark videos. The official annotation file, evaluation code, and model integration examples are maintained in the GitHub repository:
31
+
32
+ - GitHub repository: https://github.com/hrinnnn/PerceptionComp
33
+
34
+ - **Curated by:** PerceptionComp authors
35
+ - **Language(s) (NLP):** English
36
+ - **License:** Please replace `other` in the metadata above with the final data license before public release if a more specific license applies.
37
+
38
+ ### Dataset Sources
39
+
40
+ - **Repository:** https://github.com/hrinnnn/PerceptionComp
41
+ <!-- - **Paper:** Add the public paper link here when available. -->
42
+
43
+ ## Uses
44
+
45
+ ### Direct Use
46
+
47
+ PerceptionComp is intended for:
48
+
49
+ - benchmarking video-language models on complex perception-centric reasoning
50
+ - evaluating long-horizon and multi-evidence video understanding
51
+ - comparing proprietary and open-source multimodal models under a unified protocol
52
+
53
+ Users are expected to download the videos from this Hugging Face dataset and run evaluation with the official GitHub repository.
54
+
55
+ ### Out-of-Scope Use
56
+
57
+ PerceptionComp is not intended for:
58
+
59
+ - unrestricted commercial redistribution of hosted videos when original source terms do not allow it
60
+ - surveillance, identity inference, or sensitive attribute prediction
61
+ - modifying the benchmark protocol and reporting those results as directly comparable official scores
62
+
63
+ ## Dataset Structure
64
+
65
+ ### Data Instances
66
+
67
+ Each benchmark question is associated with:
68
+
69
+ - one `video_id`
70
+ - one multiple-choice question
71
+ - five answer options
72
+ - one correct answer
73
+ - one semantic category
74
+ - one difficulty label
75
+
76
+ The official annotation file is maintained in the GitHub repository:
77
+
78
+ - `benchmark/annotations/1-1114.json`
79
+
80
+ Core fields in each annotation item:
81
+
82
+ - `key`: question identifier
83
+ - `video_id`: video filename stem without `.mp4`
84
+ - `question`: question text
85
+ - `answer_choice_0` to `answer_choice_4`: five answer options
86
+ - `answer_id`: zero-based index of the correct option
87
+ - `answer`: text form of the correct answer
88
+ - `category`: semantic category
89
+ - `difficulty`: difficulty label
90
+
91
+ ### Data Files
92
+
93
+ The Hugging Face dataset stores the benchmark videos. The official evaluation code prepares them into the following local layout:
94
+
95
+ ```text
96
+ benchmark/videos/<video_id>.mp4
97
+ ```
98
+
99
+ Use the official download script from the GitHub repository:
100
+
101
+ ```bash
102
+ git clone https://github.com/hrinnnn/PerceptionComp.git
103
+ cd PerceptionComp
104
+ pip install -r requirements.txt
105
+ python scripts/download_data.py --repo-id hrinnnn/PerceptionComp
106
+ ```
107
+
108
+ ### Data Splits
109
+
110
+ The current public release uses one official evaluation set:
111
+
112
+ - `1-1114.json`: 1,114 multiple-choice questions over 273 videos
113
+
114
+ ## Dataset Creation
115
+
116
+ ### Curation Rationale
117
+
118
+ PerceptionComp was created to evaluate a failure mode that is not well covered by simpler video benchmarks: questions that require models to combine multiple perceptual constraints over time instead of relying on a single salient frame or a short summary.
119
+
120
+ ### Source Data
121
+
122
+ The benchmark uses real-world videos paired with manually written multiple-choice questions.
123
+
124
+ #### Data Collection and Processing
125
+
126
+ Videos were collected and organized for benchmark evaluation. Annotation authors then wrote perception-centric multiple-choice questions for the selected videos. Each question was designed to require visual evidence from the video rather than simple prior knowledge or caption-level shortcuts.
127
+
128
+ The release process includes:
129
+
130
+ - associating each question with a `video_id`
131
+ - formatting each sample as a five-choice multiple-choice item
132
+ - assigning semantic categories
133
+ - assigning difficulty labels
134
+ - consolidating the release into one official annotation file
135
+
136
+ #### Who are the source data producers?
137
+
138
+ The underlying videos may originate from third-party public sources. The benchmark annotations were created by the PerceptionComp authors and collaborators.
139
+
140
+ ### Annotations
141
+
142
+ #### Annotation Process
143
+
144
+ PerceptionComp contains 1,114 manually annotated five-choice questions. Questions were written to test perception-centric reasoning over videos rather than single-frame recognition alone.
145
+
146
+ #### Who are the annotators?
147
+
148
+ The annotations were created by the PerceptionComp project team.
149
+
150
+ #### Personal and Sensitive Information
151
+
152
+ The videos may contain people, faces, voices, public scenes, or other naturally occurring visual content. The dataset is intended for research evaluation, not for identity inference or sensitive attribute prediction.
153
+
154
+ ### Recommendations
155
+
156
+ Users should:
157
+
158
+ - report results with the official evaluation code
159
+ - avoid changing prompts, parsing rules, or metrics when claiming benchmark numbers
160
+ - verify that their usage complies with the terms of the original video sources
161
+ - avoid using the dataset for surveillance, identity recognition, or sensitive attribute inference
162
+
163
+ ## Citation
164
+
165
+ If you use PerceptionComp, please cite the project paper when it is publicly available.
166
+
167
+ ```bibtex
168
+ @misc{perceptioncomp2026,
169
+ title={PerceptionComp},
170
+ author={PerceptionComp Authors},
171
+ year={2026},
172
+ howpublished={Hugging Face dataset and GitHub repository}
173
+ }
174
+ ```
175
+
176
+ ## More Information
177
+
178
+ Official evaluation code and documentation:
179
+
180
+ - GitHub: https://github.com/hrinnnn/PerceptionComp
181
+
182
+ Example evaluation workflow:
183
+
184
+ ```bash
185
+ git clone https://github.com/hrinnnn/PerceptionComp.git
186
+ cd PerceptionComp
187
+ pip install -r requirements.txt
188
+ python scripts/download_data.py --repo-id hrinnnn/PerceptionComp
189
+ python evaluate/evaluate.py \
190
+ --model YOUR_MODEL_NAME \
191
+ --provider api \
192
+ --api-key YOUR_API_KEY \
193
+ --base-url YOUR_BASE_URL \
194
+ --video-dir benchmark/videos
195
+ ```
196
+
197
+ ## Dataset Card Authors
198
+
199
+ PerceptionComp authors
200
+
201
+ ## Dataset Card Contact
202
+
203
+ Add the project contact email here before publishing.