Datasets:
README upt
Browse files
README.md
CHANGED
|
@@ -2,23 +2,22 @@
|
|
| 2 |
|
| 3 |
MoSu (Most Replayed Multimodal Video Summarization) is the first large-scale multimodal video summarization dataset. It provides synchronized visual, audio, and text features for 52,678 in-the-wild videos. The ground-truth annotations are based on YouTube's "Most Replayed" statistics, offering highly reliable per-frame importance scores derived from collective viewer engagement.
|
| 4 |
|
| 5 |
-
This dataset is officially introduced in the ICLR 2026 paper, "TripleSumm: Adaptive Triple-Modality Fusion for Video Summarization".
|
| 6 |
* **Paper:** [TripleSumm: Adaptive Triple-Modality Fusion for Video Summarization](https://arxiv.org/abs/2603.01169)
|
| 7 |
* **GitHub Repository:** [smkim37/TripleSumm](https://github.com/smkim37/TripleSumm)
|
| 8 |
|
| 9 |
## Dataset Structure
|
| 10 |
|
| 11 |
-
The dataset consists of 6 core files
|
| 12 |
|
| 13 |
### 1. Metadata (`mosu_metadata.csv`)
|
| 14 |
|
| 15 |
Contains the foundational information for all 52,678 videos.
|
| 16 |
|
| 17 |
-
* **video_id**: The unique identifier for the video. This serves as the universal key to access data in all
|
| 18 |
-
* **youtube_id**: The original YouTube video ID. The video can be accessed via
|
| 19 |
* **duration**: The length of the video in seconds.
|
| 20 |
* **views**: The total view count of the video.
|
| 21 |
-
* **labels**: Original multi-label annotations provided by the YouTube-8M dataset.
|
| 22 |
* **cluster_id**: One of 10 semantic clusters (0-9). These clusters were generated based on metadata to group videos by topic (e.g., Video Games, Sports) and ensure a balanced distribution across dataset splits.
|
| 23 |
|
| 24 |
### 2. Multimodal Features (`.h5 files`)
|
|
@@ -34,10 +33,8 @@ All features have a shape of (N, D), where N corresponds to the video's duration
|
|
| 34 |
|
| 35 |
### 3. Ground Truth (`mosu_gt.h5`)
|
| 36 |
|
| 37 |
-
An HDF5 file containing the summarization labels.
|
| 38 |
|
| 39 |
-
* **Total Keys:** 52,678
|
| 40 |
-
* **Data Structure:** Each **video_id** (e.g., '005O') maps to an HDF5 Group containing four specific keys:
|
| 41 |
* **change_points**: Temporal boundaries for video shots.
|
| 42 |
* **cluster_id**: The semantic cluster ID of the video.
|
| 43 |
* **gt_score**: Frame-level ground-truth importance scores.
|
|
@@ -45,10 +42,11 @@ An HDF5 file containing the summarization labels.
|
|
| 45 |
|
| 46 |
### 4. Dataset Splits (`mosu_split.json`)
|
| 47 |
|
| 48 |
-
Contains
|
| 49 |
-
|
| 50 |
-
* **
|
| 51 |
-
* **
|
|
|
|
| 52 |
|
| 53 |
## Citation
|
| 54 |
|
|
|
|
| 2 |
|
| 3 |
MoSu (Most Replayed Multimodal Video Summarization) is the first large-scale multimodal video summarization dataset. It provides synchronized visual, audio, and text features for 52,678 in-the-wild videos. The ground-truth annotations are based on YouTube's "Most Replayed" statistics, offering highly reliable per-frame importance scores derived from collective viewer engagement.
|
| 4 |
|
|
|
|
| 5 |
* **Paper:** [TripleSumm: Adaptive Triple-Modality Fusion for Video Summarization](https://arxiv.org/abs/2603.01169)
|
| 6 |
* **GitHub Repository:** [smkim37/TripleSumm](https://github.com/smkim37/TripleSumm)
|
| 7 |
|
| 8 |
## Dataset Structure
|
| 9 |
|
| 10 |
+
The dataset consists of 6 core files providing metadata, multimodal features, ground truth annotations, and evaluation splits.
|
| 11 |
|
| 12 |
### 1. Metadata (`mosu_metadata.csv`)
|
| 13 |
|
| 14 |
Contains the foundational information for all 52,678 videos.
|
| 15 |
|
| 16 |
+
* **video_id**: The unique identifier for the video. This serves as the universal key to access data in all .h5 files and the split JSON.
|
| 17 |
+
* **youtube_id**: The original YouTube video ID. The video can be accessed via `https://www.youtube.com/watch?v={youtube_id}`.
|
| 18 |
* **duration**: The length of the video in seconds.
|
| 19 |
* **views**: The total view count of the video.
|
| 20 |
+
* **labels**: Original multi-label annotations provided by the [YouTube-8M dataset](https://research.google.com/youtube8m/).
|
| 21 |
* **cluster_id**: One of 10 semantic clusters (0-9). These clusters were generated based on metadata to group videos by topic (e.g., Video Games, Sports) and ensure a balanced distribution across dataset splits.
|
| 22 |
|
| 23 |
### 2. Multimodal Features (`.h5 files`)
|
|
|
|
| 33 |
|
| 34 |
### 3. Ground Truth (`mosu_gt.h5`)
|
| 35 |
|
| 36 |
+
An HDF5 file containing the summarization labels for all 52,678 videos. Each `video_id` (e.g., '005O') maps to an HDF5 Group containing four specific keys:
|
| 37 |
|
|
|
|
|
|
|
| 38 |
* **change_points**: Temporal boundaries for video shots.
|
| 39 |
* **cluster_id**: The semantic cluster ID of the video.
|
| 40 |
* **gt_score**: Frame-level ground-truth importance scores.
|
|
|
|
| 42 |
|
| 43 |
### 4. Dataset Splits (`mosu_split.json`)
|
| 44 |
|
| 45 |
+
Contains standardized splits for training, validation, and testing. The split ratio strictly maintains the proportional representation of each `cluster_id` for a balanced evaluation.
|
| 46 |
+
|
| 47 |
+
* **train_keys**: List of video IDs for 42,152 training videos.
|
| 48 |
+
* **val_keys**: List of video IDs for 5,263 validation videos.
|
| 49 |
+
* **test_keys**: List of video IDs for 5,263 testing videos.
|
| 50 |
|
| 51 |
## Citation
|
| 52 |
|