Datasets:
File size: 4,846 Bytes
2e9afd8 ab96ef4 2e9afd8 ab96ef4 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 | ---
license: mit
language:
- en
pretty_name: LongTVQA+
---
# LongTVQA+ Dataset
This repository contains the **LongTVQA+** dataset in JSON format.
LongTVQA+ is built upon the original **TVQA+** dataset, with the key difference that it **extends the question grounding scope from short clip-level segments (≈1 minute) to long episode-level videos (up to ~20 minutes)**.
This enables research on long-form video understanding, long-range temporal reasoning, and fine-grained spatio-temporal grounding in realistic TV show episodes.
In addition to the extended temporal scope, LongTVQA+ preserves and leverages the rich annotations provided in TVQA+, including:
1. Frame-level bounding box annotations for visual concept words appearing in questions and correct answers.
2. Refined timestamp annotations aligned with long episode-level context.
Please refer to the original **TVQA+ paper** for details on the annotation protocol and baseline evaluations.
---
## Files
- `LongTVQA_plus_train.json` — training split (23,545 QA samples)
- `LongTVQA_plus_val.json` — validation split (3,017 QA samples)
- `LongTVQA_plus_subtitle_clip_level.json` — clip-level subtitles indexed by video clip (4,198 clips)
- `LongTVQA_plus_subtitle_episode_level.json` — episode-level subtitles indexed by episode (220 episodes)
---
## QA JSON Format
Each entry in `LongTVQA_plus_train.json` and `LongTVQA_plus_val.json` is a dictionary with the following fields:
| Key | Type | Description |
| --- | --- | --- |
| `qid` | int | Question ID (same as in TVQA+). |
| `q` | str | Question text. |
| `a0` ... `a4` | str | Five multiple-choice answers. |
| `answer` | str | Correct answer key (`"a0"`–`"a4"`). |
| `ts` | list | Refined timestamp annotation. For example, `[0, 5.4]` indicates the localized temporal span starts at 0s and ends at 5.4s. |
| `episode_name` | str | Episode ID (e.g. `s01e02`). |
| `occur_clip` | str | Video clip name. Format: `{show_name_abbr}_s{season}e{episode}_seg{segment}_clip_{clip}`. Episodes are typically divided into two segments separated by the opening theme. For **The Big Bang Theory**, `{show_name_abbr}` is omitted (e.g. `s05e02_seg02_clip_00`). |
| `bbox` | dict | Frame-level bounding box annotations sampled at 3 FPS. Keys are frame indices. Values are lists of bounding boxes with `img_id`, `top`, `left`, `width`, `height`, and `label`. |
---
### QA Sample
```json
{
"answer": "a1",
"qid": 134094,
"ts": [5.99, 11.98],
"a1": "Howard is talking to Raj and Leonard",
"a0": "Howard is talking to Bernadette",
"a3": "Howard is talking to Leonard and Penny",
"a2": "Howard is talking to Sheldon , and Raj",
"q": "Who is Howard talking to when he is in the lab room ?",
"episode_name": "s05e02",
"occur_clip": "s05e02_seg02_clip_00",
"a4": "Howard is talking to Penny and Bernadette",
"bbox": {
"14": [
{
"img_id": 14,
"top": 153,
"label": "Howard",
"width": 180,
"height": 207,
"left": 339
},
{
"img_id": 14,
"top": 6,
"label": "lab",
"width": 637,
"height": 354,
"left": 3
}
],
"20": [],
"26": [],
"32": [],
"38": []
}
}
```
---
## Subtitles JSON Format
Two subtitle files are provided to support different temporal granularities:
| File | Key | Type | Description |
| ------------------------------------------- | -------------- | ---- | ------------------------------------------------------------------------------------------------------------------- |
| `LongTVQA_plus_subtitle_clip_level.json` | `vid_name` | str | Clip-level subtitle text, with utterances separated by `<eos>`. |
| `LongTVQA_plus_subtitle_episode_level.json` | `episode_name` | str | Episode-level subtitle text, including clip markers such as `<seg01_clip_00>`, and utterances separated by `<eos>`. |
---
### Subtitles Sample
```json
{
"s09e14_seg02_clip_04": "Sheldon : That 's a risk I'm willing to take ! <eos> Amy : Well , this is so nice . <eos> ..."
}
```
---
## License
This dataset is released under the **MIT License**.
## 📝 Citation
If you find our work helpful, please cite:
```bibtex
@misc{liu2025longvideoagentmultiagentreasoninglong,
title={LongVideoAgent: Multi-Agent Reasoning with Long Videos},
author={Runtao Liu and Ziyi Liu and Jiaqi Tang and Yue Ma and Renjie Pi and Jipeng Zhang and Qifeng Chen},
year={2025},
eprint={2512.20618},
archivePrefix={arXiv},
primaryClass={cs.AI},
url={[https://arxiv.org/abs/2512.20618](https://arxiv.org/abs/2512.20618)},
} |