Datasets:

Modalities:
Text
Video
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
License:
ZhangYuanhan commited on
Commit
ed825f9
·
verified ·
1 Parent(s): d6e3062

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -10
README.md CHANGED
@@ -58,31 +58,94 @@ configs:
58
  path: data/test_primary_oe-*
59
  ---
60
 
61
- # Towards Video Turing Test (Video-TT): Video Comprehension and Reasoning Benchmark with Complex Visual Narratives
62
 
63
- Video-TT comprises 1,000 YouTube Shorts videos, each with one open-ended question and four adversarial questions that probe visual and narrative complexity.
64
 
65
  ## 🚀 What's New
66
- - **\[2025.03\]** We release the benchmark!
67
 
68
- ## 1. Why we need a new benchmark like Video-TT?
69
 
70
- - **Sampling v.s Understanding:** Current video understanding benchmark do not clearly distinguish between errors caused by insufficient frame sampling and errors caused by failures in actual video understanding. We ensure that each question in Video-TT can be well answered with uniformly sampled 80 frames where most current video model can easily handle.v
71
- - **Pursing human level video understanding:** We carefully select Q&A, where human can achieve 87.5 points, but even the best model can only achieve 47.5. In comparison, without sampling issue, current video model can achieve more than 85 points on video understanding benchmark [1]
72
 
73
- ## 2. How to ensure the quality of Video-TT?
74
 
 
75
 
76
- ## 3. How to evaluate on Video-TT
 
 
 
 
 
 
 
77
 
78
- ## 4. Dataset summary
 
79
 
 
80
 
 
81
 
82
- [1] Fu, Chaoyou, et al. "Video-mme: The first-ever comprehensive evaluation benchmark of multi-modal llms in video analysis." arXiv preprint arXiv:2405.21075 (2024).
83
 
 
84
 
 
 
 
 
 
 
85
 
 
86
 
 
87
 
 
88
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  path: data/test_primary_oe-*
59
  ---
60
 
61
+ # Towards Video Turing Test (Video-TT): A Benchmark for Video Comprehension and Reasoning with Complex Visual Narratives
62
 
63
+ Video-TT comprises 1,000 YouTube videos, each paired with one open-ended question and four adversarial questions designed to probe visual and narrative complexity.
64
 
65
  ## 🚀 What's New
66
+ - **[2025.03]** We release the benchmark!
67
 
68
+ ## 1. Why Do We Need a New Benchmark Like Video-TT?
69
 
70
+ - **Sampling vs. Understanding:** Current video understanding benchmarks do not clearly distinguish between errors caused by insufficient frame sampling and errors due to failures in actual video comprehension. We ensure that each question in Video-TT can be answered with 80 uniformly sampled frames—frames that most current video models can easily process.
71
+ - **Pursuing Human-Level Video Understanding:** We carefully select Q&A pairs where humans achieve an 87.5% accuracy rate, while the best models only reach 47.5%. In contrast, when sampling is not a limiting factor, current video models score above 85% on existing video understanding benchmarks [1].
72
 
73
+ ## 2. Dataset Summary
74
 
75
+ We introduce five question prototypes.
76
 
77
+ ### Dataset Statistics
78
+ - **Number of videos:** 1,000, primarily in formats like YouTube Shorts.
79
+ - **Number of Q&A pairs:** Each video is paired with five questions that probe different aspects of the same content, designed to challenge model robustness in adversarial scenarios:
80
+ - Primary open-ended question
81
+ - Paraphrased open-ended question
82
+ - Correctly-led open-ended question
83
+ - Wrongly-led open-ended question
84
+ - Multiple-choice question
85
 
86
+ One example of five question types is shown below:
87
+ <img src="dataset_expansion_page-0001.jpg" style="width:30%;">
88
 
89
+ ### Evaluation Metrics
90
 
91
+ #### Correctness:
92
 
93
+ Measures accuracy for each question type. We use Qwen2.5-72B as the judge for open-ended questions and a rule-based method for multiple-choice questions.
94
 
95
+ #### Robustness:
96
 
97
+ Let:
98
+ - A_primary_correct be the set of videos where the primary open-ended question is answered correctly.
99
+ - A_paraphrased_correct be the set of videos where the paraphrased open-ended question is answered correctly.
100
+ - A_correctly_led_correct be the set of videos where the correctly-led open-ended question is answered correctly.
101
+ - A_wrongly_led_correct be the set of videos where the wrongly-led open-ended question is answered correctly.
102
+ - A_multiple_choice_correct be the set of videos where the multiple-choice question is answered correctly.
103
 
104
+ The set of videos where all five questions are answered correctly, denoted as A_full_correct, is the intersection of all these sets:
105
 
106
+ A_full_correct = A_primary_correct ∩ A_paraphrased_correct ∩ A_correctly_led_correct ∩ A_wrongly_led_correct ∩ A_multiple_choice_correct
107
 
108
+ Thus, the **Robustness Score (RB)** becomes:
109
 
110
+ R = |A_full_correct| / |A_primary_correct|
111
+
112
+ Where |A| denotes the size of the set A, representing the number of videos in that set.
113
+
114
+
115
+ ## 3. Ensuring the Quality of Video-TT
116
+
117
+ <img src="annotation_pipeline_page-0001.jpg" style="width:50%;">
118
+
119
+ The Video-TT annotation process consists of four stages:
120
+
121
+ - **Stage 1:** We select complex videos that allow for challenging questions. (Refer to the paper for the definition of complexity factors.)
122
+ - **Stage 2:** We prompt several vision-language models (VLMs) to check whether a question can be answered. If a model answers it correctly, we remove the question.
123
+ - **Stage 3:** For each remaining question, we provide an answer and an explanation.
124
+ - **Stage 4:** We manually verify that each question can be answered using 80 uniformly sampled frames.
125
+
126
+ For more details, please refer to the paper.
127
+
128
+ ## 4. How to Evaluate on Video-TT
129
+
130
+ (To be detailed in the paper.)
131
+
132
+ ## 5. Leaderboard
133
+
134
+ For the latest leaderboard, please refer to **XXX**. You can submit your evaluations there. Some results are produced by us, while others are submitted by external researchers.
135
+
136
+ To reproduce our results, please check **XXX** for evaluation scripts.
137
+
138
+ <img src="771742752771_.pic.jpg" style="width:50%;">
139
+
140
+ ## 6. Dataset Maintenance
141
+
142
+ If you find any mistakes in the dataset, please submit the corresponding `question_id` to our issue page. Our team is committed to maintaining this dataset in the long run to ensure its quality.
143
+
144
+ ## 7. Acknowledgments
145
+
146
+ (Include acknowledgments as needed.)
147
+
148
+
149
+
150
+
151
+ [1] Fu, Chaoyou, et al. "Video-mme: The first-ever comprehensive evaluation benchmark of multi-modal llms in video analysis." arXiv preprint arXiv:2405.21075 (2024).