Update README.md
Browse files
README.md
CHANGED
|
@@ -19,4 +19,36 @@ tags:
|
|
| 19 |
pretty_name: StoryFrames
|
| 20 |
size_categories:
|
| 21 |
- 1K<n<10K
|
| 22 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
pretty_name: StoryFrames
|
| 20 |
size_categories:
|
| 21 |
- 1K<n<10K
|
| 22 |
+
---
|
| 23 |
+
|
| 24 |
+
# The StoryFrames Dataset
|
| 25 |
+
StoryFrames is a human-annotated dataset created to enhance a model's capability of understanding and reasoning over sequences of images.
|
| 26 |
+
It is specifically designed for tasks like generating a description for the next scene in a story based on previous visual and textual information.
|
| 27 |
+
The dataset repurposes the [StoryBench dataset](https://arxiv.org/abs/2308.11606), a video dataset originally designed to predict future frames of a video.
|
| 28 |
+
StoryFrames subsamples frames from those videos and pairs them with annotations for the task of _next-description prediction_.
|
| 29 |
+
Each "story" is a sample of the dataset and can vary in length and complexity.
|
| 30 |
+
|
| 31 |
+
The dataset contains 8,881 samples, divided into train and validation splits.
|
| 32 |
+

|
| 33 |
+
|
| 34 |
+
## What Is a Story in StoryFrames?
|
| 35 |
+
* **A story is a sequence of scenes:**
|
| 36 |
+
Each story is composed of multiple scenes, where each scene is a part of the overall narrative.
|
| 37 |
+
|
| 38 |
+
* **Scenes consist of two main components:**
|
| 39 |
+
* **Images**: Each scene is made up of several frames (images) that have been subsampled from the original video.
|
| 40 |
+
* **Scene Description**: There is a single textual description for each scene (i.e., one or more images) that captures the plot of the scene.
|
| 41 |
+
|
| 42 |
+
## How Is the Data Organized?
|
| 43 |
+
* **Temporal Markers:**
|
| 44 |
+
* `start_times` and `end_times`: These fields provide the time markers indicating when each scene begins and ends in the video. They define the boundaries of each scene.
|
| 45 |
+
|
| 46 |
+
* **Frame Subsampling:**
|
| 47 |
+
* `subsampled_frames_per_scene`: For each scene, a list of frame timestamps is provided. Each timestamp is formatted to show the second and millisecond (for example, `frame_sec.millisec` would be `frame_1.448629`). These timestamps indicate which frames were selected from the scene.
|
| 48 |
+
|
| 49 |
+
* **Image Data:**
|
| 50 |
+
* `scenes`: In a structure that mirrors the subsampled timestamps, this field contains the actual images that were extracted. The images are organized as a list of lists: each inner list corresponds to one scene and contains the images in the order they were sampled.
|
| 51 |
+
|
| 52 |
+
* **Narrative Descriptions:**
|
| 53 |
+
* `sentence_parts`: This field contains a list of strings. Each string provides a description for one scene in the story. Even though a scene is made up of multiple images, the corresponding description captures the plot progression over all images of that scene.
|
| 54 |
+
|