Files changed (1) hide show
  1. README.md +150 -36
README.md CHANGED
@@ -3,80 +3,194 @@ license: cc-by-nc-4.0
3
  task_categories:
4
  - video-text-to-text
5
  - visual-question-answering
6
- language: [en]
7
- tags: [video, temporal-grounding, one-to-many, benchmark, evaluation, multimodal, mllm]
 
 
 
 
 
 
 
 
8
  pretty_name: OMTG Bench
9
- size_categories: [n<1K]
 
10
  ---
11
 
12
  # OMTG Bench: A Benchmark for One-to-Many Temporal Grounding
13
 
14
- OMTG Bench is the first comprehensive benchmark tailored for One-to-Many Temporal Grounding,
15
- introduced in **"Towards One-to-Many Temporal Grounding"** (ICML 2026, under review).
 
16
 
17
  ## Dataset Summary
18
 
19
  | Item | Value |
20
  |---|---|
 
21
  | # Samples | 340 |
22
- | Annotation | Manually curated, expert-verified |
23
- | Consistency | > 90% |
24
- | Source video pools | Charades, ActivityNet, QVHighlights, VTimeLLM, Moment-10M (test only) |
25
- | Overlap with training | None |
26
- | Domains | Sports, cooking, news, etc. |
 
 
 
27
 
28
- ### Distribution
29
  | Property | Value |
30
  |---|---|
31
- | GT segments per query | 2 – 20 |
32
- | 2–3 segments | 62.2% |
33
- | > 6 segments | 15% |
34
- | Video duration | 21 s – 17 min (avg 221.6 s) |
 
 
 
 
 
 
 
35
 
36
  ## Evaluation Metrics
 
 
 
37
  | Metric | Description |
38
  |---|---|
39
- | tIoU | Average temporal IoU over unions |
40
- | C-Acc | Count Accuracy (cardinality correctness) |
41
- | tF1@ξ | Temporal F1 at IoU ξ ∈ {0.3, 0.5, 0.7} via optimal bipartite matching |
42
- | **EtF1** | Effective Temporal F1 — strictly penalizes wrong cardinality (primary metric) |
 
 
 
 
 
 
43
 
44
- ## Baseline Results (OMTG Bench)
45
  | Model | C-Acc | tF1@0.3 | tF1@0.5 | tF1@0.7 | tIoU | EtF1 |
46
  |---|---|---|---|---|---|---|
47
- | Qwen2.5-VL-3B | 0.00 | 15.17 | 7.01 | 2.86 | 11.60 | |
48
- | Gemini-3-Pro | 30.63 | 58.30 | 47.75 | 29.89 | 47.63 | 21.30 |
49
- | Gemini-2.5-Pro | 50.94 | 55.72 | 43.57 | 27.97 | 43.24 | 27.80 |
50
  | Seed-1.8 | 38.12 | 67.13 | 54.67 | 38.79 | 56.81 | 28.04 |
51
- | **OMTG-4B (ours)** | | | | | | **43.65** |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
 
53
  ## Dataset Structure
 
 
 
54
  ```
55
  omtg_bench/
56
- ├── OMTGBench.tsv # 340 samples
57
- └── videos.zip # ≈3.74 GB
58
  ```
59
 
60
- ### Data Fields
 
 
 
61
  | Field | Type | Description |
62
  |---|---|---|
63
- | video_id | string | Video identifier |
64
- | duration | float | Video length (s) |
65
- | query | string | Natural-language query |
66
- | segments | list[[float,float]] | GT `[start,end]` intervals |
67
- | source | string | Charades / ActivityNet / QVHighlights / VTimeLLM / Moment-10M |
68
- | domain | string | Topical domain |
69
 
70
  ## Usage
 
71
  ```python
72
  from datasets import load_dataset
 
73
  bench = load_dataset("insomnia7/omtg_bench", split="train")
 
74
  ```
75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
  ## Companion Dataset
77
- Training dataset: [insomnia7/omtg56k](https://huggingface.co/datasets/insomnia7/omtg56k) (46k SFT + 10k RL).
 
78
 
79
  ## License
80
- CC BY-NC 4.0 — non-commercial research only.
81
 
82
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  task_categories:
4
  - video-text-to-text
5
  - visual-question-answering
6
+ language:
7
+ - en
8
+ tags:
9
+ - video
10
+ - temporal-grounding
11
+ - one-to-many
12
+ - benchmark
13
+ - evaluation
14
+ - multimodal
15
+ - mllm
16
  pretty_name: OMTG Bench
17
+ size_categories:
18
+ - n<1K
19
  ---
20
 
21
  # OMTG Bench: A Benchmark for One-to-Many Temporal Grounding
22
 
23
+ **OMTG Bench** is the first comprehensive benchmark tailored for the **One-to-Many Temporal Grounding (OMTG)** task, introduced in the paper **"Towards One-to-Many Temporal Grounding"** (ICML 2026, under review).
24
+
25
+ Unlike conventional temporal grounding benchmarks that assume a one-to-one mapping between a query and a single temporal segment, OMTG Bench evaluates a model's ability to retrieve **all** disjoint temporal segments in a video that correspond to a single textual query — a setting that is ubiquitous in real-world video content but on which state-of-the-art MLLMs (including Gemini-2.5-Pro, Gemini-3-Pro, Seed-1.8, and Qwen2.5-VL / Qwen3-VL families) exhibit a critical capability gap.
26
 
27
  ## Dataset Summary
28
 
29
  | Item | Value |
30
  |---|---|
31
+ | Task | One-to-Many Temporal Grounding (OMTG) |
32
  | # Samples | 340 |
33
+ | Annotation | Manually curated & verified by independent experts |
34
+ | Inter-annotator consistency | > 90% |
35
+ | Source video pools | Charades, ActivityNet, QVHighlights, VTimeLLM, Moment-10M (test splits only) |
36
+ | Overlap with training set | None (by construction) |
37
+ | Domains | Sports, cooking, news, and other diverse real-world scenarios |
38
+ | Language | English |
39
+
40
+ ### Key Distribution Statistics
41
 
 
42
  | Property | Value |
43
  |---|---|
44
+ | Ground-truth segments per query | 2 – 20 |
45
+ | 2–3 segments per query | 62.2% of samples |
46
+ | > 6 segments per query | 15% of samples |
47
+ | Video duration | 21 s – 17 min |
48
+ | Average video duration | 221.6 s |
49
+
50
+ This distribution makes OMTG Bench a severe test of both **event cardinality perception** and **fine-grained temporal localization**, covering short clips as well as long-form narratives.
51
+
52
+ ## Supported Tasks
53
+
54
+ - **One-to-Many Temporal Grounding** — given a video `V` and a textual query `Q`, predict the set `{[s₁, e₁], [s₂, e₂], …, [sₖ, eₖ]}` of all disjoint intervals where `Q` occurs.
55
 
56
  ## Evaluation Metrics
57
+
58
+ The benchmark introduces a rigorous metric suite that jointly measures precision, recall, and event-count correctness:
59
+
60
  | Metric | Description |
61
  |---|---|
62
+ | **tIoU** | Average temporal IoU over unions of predicted and ground-truth intervals |
63
+ | **C-Acc** | Count Accuracy — the percentage of samples whose predicted cardinality equals the ground-truth cardinality |
64
+ | **tF1@ξ** | Temporal F1-Score at IoU threshold `ξ ∈ {0.3, 0.5, 0.7}` under optimal bipartite matching |
65
+ | **EtF1** | **Effective Temporal F1**assigns zero credit to any sample with incorrect predicted cardinality, strictly penalizing under- and over-retrieval and hallucinations |
66
+
67
+ EtF1 is the recommended primary metric, as it couples instance-level precision–recall with event-count correctness.
68
+
69
+ ## Baseline Results on OMTG Bench
70
+
71
+ A comprehensive assessment of representative open-source and proprietary MLLMs. All numbers are percentages (%).
72
 
 
73
  | Model | C-Acc | tF1@0.3 | tF1@0.5 | tF1@0.7 | tIoU | EtF1 |
74
  |---|---|---|---|---|---|---|
75
+ | **Proprietary Models** | | | | | | |
 
 
76
  | Seed-1.8 | 38.12 | 67.13 | 54.67 | 38.79 | 56.81 | 28.04 |
77
+ | Gemini-2.5-Pro | 50.94 | 55.72 | 43.57 | 27.97 | 43.24 | 27.80 |
78
+ | Gemini-3-Pro | 30.63 | 58.30 | 47.75 | 29.89 | 47.63 | 21.30 |
79
+ | **Open-Source General MLLMs** | | | | | | |
80
+ | Qwen2.5-VL-3B | 0.00 | 15.17 | 7.01 | 2.86 | 11.60 | 0.00 |
81
+ | Qwen2.5-VL-7B | 0.00 | 21.04 | 12.08 | 7.14 | 20.35 | 0.00 |
82
+ | Qwen2.5-VL-32B | 0.00 | 16.81 | 9.66 | 4.76 | 18.32 | 0.00 |
83
+ | Qwen2.5-VL-72B | 0.00 | 21.16 | 12.20 | 6.88 | 20.02 | 0.00 |
84
+ | Qwen3-VL-4B | 0.31 | 37.07 | 26.75 | 17.93 | 30.42 | 0.21 |
85
+ | Qwen3-VL-8B | 0.00 | 37.73 | 27.02 | 18.70 | 30.62 | 0.00 |
86
+ | Qwen3-VL-30B | 0.00 | 37.03 | 25.98 | 17.52 | 32.36 | 0.00 |
87
+ | Qwen3-VL-235B | 0.31 | 34.66 | 25.25 | 16.45 | 25.56 | 0.21 |
88
+ | **Temporal-Grounding Experts** | | | | | | |
89
+ | VideoChat-R1-7B | 0.00 | 32.07 | 19.70 | 10.42 | 24.93 | 0.00 |
90
+ | VideoChat-R1.5-7B | 0.31 | 28.41 | 15.53 | 9.85 | 27.96 | 0.10 |
91
+ | Time-R1-7B | 0.00 | 28.94 | 18.73 | 10.00 | 24.11 | 0.00 |
92
+ | UniTime | 0.00 | 35.27 | 30.15 | 23.58 | 37.12 | 0.00 |
93
+ | Timelens-8B | 0.00 | 39.14 | 32.76 | 22.58 | 32.38 | 0.00 |
94
+ | **Ours** | | | | | | |
95
+ | **OMTG-4B** | **55.63** | **73.46** | **65.40** | **48.96** | **61.24** | **43.65** |
96
+
97
+ ### Performance Gain of OMTG-4B across Training Stages
98
+
99
+ | Stage | C-Acc | tF1@0.3 | tF1@0.5 | tF1@0.7 | tIoU | EtF1 |
100
+ |---|---|---|---|---|---|---|
101
+ | Base (Qwen3-VL-4B) | 0.31 | 37.07 | 26.75 | 17.93 | 30.42 | 0.21 |
102
+ | + SFT (on OMTG-56K) | 44.06 | 69.57 | 61.23 | 45.63 | 56.94 | 34.81 |
103
+ | + RL (GRPO) | **55.63** | **73.46** | **65.40** | **48.96** | **61.24** | **43.65** |
104
+
105
+ Key observations:
106
+
107
+ - Standard open-source models (Qwen2.5-VL series) yield **0% C-Acc / 0% EtF1**, failing to capture the one-to-many complexity.
108
+ - Even the newer Qwen3-VL family (4B → 235B) barely exceeds 0% on cardinality-sensitive metrics.
109
+ - Advanced proprietary models (Seed-1.8, Gemini-2.5/3-Pro) reach only **21–28% EtF1**.
110
+ - Our **OMTG-4B** achieves a new state of the art at **43.65% EtF1**, outperforming the best proprietary baseline (Seed-1.8) by **+15.61 EtF1**.
111
 
112
  ## Dataset Structure
113
+
114
+ ### Files
115
+
116
  ```
117
  omtg_bench/
118
+ ├── OMTGBench.tsv # 340 annotated samples (74.2 kB)
119
+ └── videos.zip # corresponding source videos (≈3.74 GB)
120
  ```
121
 
122
+ ### Data Fields (`OMTGBench.tsv`)
123
+
124
+ Each row is one OMTG evaluation sample. Typical columns:
125
+
126
  | Field | Type | Description |
127
  |---|---|---|
128
+ | `video_id` | string | Unique video identifier (maps to a file in `videos.zip`) |
129
+ | `duration` | float | Video length in seconds |
130
+ | `query` | string | Natural-language query describing a repeating event |
131
+ | `segments` | list[[float, float]] | Ground-truth `[start, end]` timestamps in seconds for all occurrences |
132
+ | `source` | string | Originating dataset (Charades / ActivityNet / QVHighlights / VTimeLLM / Moment-10M) |
133
+ | `domain` | string | Topical domain (e.g., sports, cooking, news) |
134
 
135
  ## Usage
136
+
137
  ```python
138
  from datasets import load_dataset
139
+
140
  bench = load_dataset("insomnia7/omtg_bench", split="train")
141
+ print(bench[0])
142
  ```
143
 
144
+ Videos can be extracted from `videos.zip` and matched via `video_id`.
145
+
146
+ ### Recommended Evaluation Protocol
147
+
148
+ 1. For each sample, prompt the model with the video and the textual query.
149
+ 2. Parse the model output into a set of `[start, end]` intervals (seconds).
150
+ 3. Compute **C-Acc**, **tF1@{0.3, 0.5, 0.7}**, **tIoU**, and **EtF1** via optimal bipartite matching against the ground-truth segments.
151
+ 4. Report **EtF1** as the primary metric.
152
+
153
+ ## Construction & Quality Control
154
+
155
+ Samples were **randomly sampled and then manually curated** from the *test* splits of Charades, ActivityNet, QVHighlights, VTimeLLM, and Moment-10M. Care was taken to ensure:
156
+
157
+ - **Zero overlap** between benchmark videos and any training data (including the companion [OMTG-56K](https://huggingface.co/datasets/insomnia7/omtg56k) dataset).
158
+ - **Precise boundary annotations** verified by independent experts.
159
+ - **Inter-annotator consistency > 90%**.
160
+
161
+ ## Intended Use
162
+
163
+ - Evaluating MLLMs on one-to-many video temporal grounding.
164
+ - Studying event cardinality perception and hallucination in video-language models.
165
+ - Diagnosing over- and under-retrieval behaviors in long-form video understanding.
166
+
167
+ ## Limitations
168
+
169
+ - Source videos inherit biases and licenses from the original datasets.
170
+ - Domain coverage, while diverse, is not exhaustive; extremely long videos (> 20 min) are underrepresented.
171
+ - Manual annotations, although expert-verified, may contain rare residual noise.
172
+ - Intended for non-commercial research only.
173
+
174
  ## Companion Dataset
175
+
176
+ - Training dataset: [insomnia7/omtg56k](https://huggingface.co/datasets/insomnia7/omtg56k) — 56k high-quality instruction-tuning samples (46k SFT + 10k RL).
177
 
178
  ## License
 
179
 
180
+ Released under **CC BY-NC 4.0** for non-commercial research use. Source videos remain subject to their original dataset licenses.
181
+
182
+ ## Citation
183
+
184
+ ```bibtex
185
+ @inproceedings{omtg2026,
186
+ title = {Towards One-to-Many Temporal Grounding},
187
+ author = {Anonymous Authors},
188
+ booktitle = {Proceedings of the International Conference on Machine Learning (ICML)},
189
+ year = {2026},
190
+ note = {Under review}
191
+ }
192
+ ```
193
+
194
+ ## Contact
195
+
196
+ For questions or issues, please open a discussion on the dataset's Hugging Face page.