Hokin commited on
Commit
d19ef21
·
verified ·
1 Parent(s): a5ad5c9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +142 -0
README.md ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: derivative-mixed
4
+ license_link: LICENSE
5
+ task_categories:
6
+ - visual-question-answering
7
+ - video-classification
8
+ tags:
9
+ - video
10
+ - mcqa
11
+ - vqa
12
+ - video-generation
13
+ - wan2.2
14
+ - i2v
15
+ - vbvr
16
+ size_categories:
17
+ - 1K<n<10K
18
+ ---
19
+
20
+ # Video-MCP
21
+
22
+ **Video-MCP** is a synthetic video dataset for training and evaluating video generation models on **multiple-choice question-answering (MCQA)** tasks. Each sample is a short video clip (~5 seconds) where a visual question-answering prompt is embedded directly into the video frames, and the correct answer is revealed by progressively highlighting one of four answer boxes (A/B/C/D) over the duration of the clip.
23
+
24
+ The dataset is designed for fine-tuning image-to-video models (specifically **Wan2.2-I2V-A14B**) to produce videos that "answer" visual questions by highlighting the correct option.
25
+
26
+ Output follows the **[VBVR DataFactory](https://github.com/video-reason/VBVR-DataFactory)** directory convention.
27
+
28
+ ## Dataset Details
29
+
30
+ | Property | Value |
31
+ |---|---|
32
+ | **Version** | 1.0 |
33
+ | **Total samples** | 6,912 |
34
+ | **Video resolution** | 832x480 |
35
+ | **Frame count** | 81 frames per clip |
36
+ | **Frame rate** | 16 FPS |
37
+ | **Duration** | ~5.06 seconds per clip |
38
+ | **Codec** | H.264, yuv420p, MP4 container |
39
+ | **Highlight style** | darken (default) |
40
+
41
+ ## Source Datasets
42
+
43
+ Video-MCP draws from four publicly available MCQA-VQA datasets on Hugging Face:
44
+
45
+ | Generator ID | Name | Source | Samples | Domain |
46
+ |---|---|---|---|---|
47
+ | M-1 | corecognition | `williamium/CoreCognition` | 753 | General visual reasoning |
48
+ | M-2 | scienceqa | `derek-thomas/ScienceQA` | 3,905 | Science education (image-only subset) |
49
+ | M-3 | mathvision | `MathLLMs/MathVision` | 1,254 | Competition math with diagrams |
50
+ | M-4 | phyx | `Cloudriver/PhyX` | 1,000 | Physics reasoning |
51
+
52
+ All source datasets are filtered to include only samples that have an associated image and exactly four answer choices (A/B/C/D).
53
+
54
+ ## Data Structure
55
+
56
+ Each sample follows the [VBVR DataFactory](https://github.com/video-reason/VBVR-DataFactory) directory convention:
57
+
58
+ ```
59
+ {generator_id}_{name}_data-generator/
60
+ clip_config.json
61
+ {name}_task/
62
+ {name}_{NNNN}/
63
+ first_frame.png # Frame 0: question visible, no highlight
64
+ prompt.txt # Plain-text question, choices, and answer
65
+ final_frame.png # Last frame: correct answer fully highlighted
66
+ ground_truth.mp4 # Full clip with progressive answer reveal
67
+ original/
68
+ question.json # Structured metadata (JSON)
69
+ <source_image> # Original image from source dataset
70
+ ```
71
+
72
+ ### File Descriptions
73
+
74
+ | File | Description |
75
+ |---|---|
76
+ | `first_frame.png` | The opening frame showing the question panel (image + question text + four choices) with A/B/C/D answer boxes in the corners. No answer is highlighted. |
77
+ | `final_frame.png` | The closing frame with the correct answer box fully highlighted. |
78
+ | `ground_truth.mp4` | The complete video clip. The correct answer gradually highlights from frame 1 to the final frame (linear fade-in). |
79
+ | `prompt.txt` | Human-readable text: question, choices (A/B/C/D), and the correct answer letter. |
80
+ | `original/question.json` | Structured JSON with fields: `dataset`, `source_id`, `question`, `choices`, `answer`, `original_image_filename`. |
81
+ | `original/<image>` | The raw source image preserved with its original filename. |
82
+ | `clip_config.json` | Generator-level config: `fps`, `seconds`, `num_frames`, `width`, `height`. |
83
+
84
+ ### Frame Layout
85
+
86
+ Each frame uses a two-column layout:
87
+ - **Left column**: the source VQA image, scaled to fill.
88
+ - **Right column**: question text and the four answer options.
89
+ - **Corners**: A (top-left), B (top-right), C (bottom-left), D (bottom-right) answer boxes.
90
+
91
+ ### prompt.txt Format
92
+
93
+ ```
94
+ What color is the object in the image?
95
+
96
+ A: Red
97
+ B: Blue
98
+ C: Green
99
+ D: Yellow
100
+
101
+ Answer: A
102
+ ```
103
+
104
+ ## Video Specifications
105
+
106
+ These defaults align with **Wan2.2-I2V-A14B** fine-tuning constraints:
107
+
108
+ - **Resolution**: 832x480 (width and height divisible by 8 for VAE spatial compression)
109
+ - **Frames**: 81 (satisfies `1 + 4k` for VAE temporal grid)
110
+ - **FPS**: 16
111
+ - **Duration**: ~5.06 seconds
112
+ - **Codec**: H.264, yuv420p pixel format
113
+
114
+ ## Intended Use
115
+
116
+ - Fine-tuning image-to-video generation models to produce MCQA-answering videos
117
+ - Evaluating video generation models on structured visual reasoning tasks
118
+ - Research on embedding structured UI interactions into generated video
119
+
120
+ ## Limitations
121
+
122
+ - All source questions are filtered to exactly 4 choices (A/B/C/D); questions with fewer or more options are excluded.
123
+ - The answer highlight is a simple linear fade-in; no complex visual dynamics.
124
+ - Source images and questions inherit any biases or errors from the upstream HF datasets.
125
+ - The dataset uses a single fixed resolution (832x480) and frame count (81).
126
+
127
+ ## Citation
128
+
129
+ If you use this dataset, please cite the source datasets:
130
+
131
+ - **CoreCognition**: `williamium/CoreCognition` on Hugging Face
132
+ - **ScienceQA**: Lu et al., "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering" (NeurIPS 2022)
133
+ - **MathVision**: Wang et al., "MathVision: Measuring Multimodal Mathematical Reasoning with Benchmarks" (2024)
134
+ - **PhyX**: `Cloudriver/PhyX` on Hugging Face
135
+
136
+ ## License
137
+
138
+ This dataset is a derivative work. Each source dataset has its own license terms. Users should verify compliance with upstream licenses before redistribution.
139
+
140
+ ## Generation Code
141
+
142
+ [https://github.com/video-reason/video-mcp](https://github.com/video-reason/video-mcp)