Datasets:

Modalities:
Text
Formats:
json
Languages:
English
ArXiv:
License:
LongTVQA_plus / README.md
z1ya0's picture
Update README.md
ab96ef4 verified
---
license: mit
language:
- en
pretty_name: LongTVQA+
---
# LongTVQA+ Dataset
This repository contains the **LongTVQA+** dataset in JSON format.
LongTVQA+ is built upon the original **TVQA+** dataset, with the key difference that it **extends the question grounding scope from short clip-level segments (≈1 minute) to long episode-level videos (up to ~20 minutes)**.
This enables research on long-form video understanding, long-range temporal reasoning, and fine-grained spatio-temporal grounding in realistic TV show episodes.
In addition to the extended temporal scope, LongTVQA+ preserves and leverages the rich annotations provided in TVQA+, including:
1. Frame-level bounding box annotations for visual concept words appearing in questions and correct answers.
2. Refined timestamp annotations aligned with long episode-level context.
Please refer to the original **TVQA+ paper** for details on the annotation protocol and baseline evaluations.
---
## Files
- `LongTVQA_plus_train.json` — training split (23,545 QA samples)
- `LongTVQA_plus_val.json` — validation split (3,017 QA samples)
- `LongTVQA_plus_subtitle_clip_level.json` — clip-level subtitles indexed by video clip (4,198 clips)
- `LongTVQA_plus_subtitle_episode_level.json` — episode-level subtitles indexed by episode (220 episodes)
---
## QA JSON Format
Each entry in `LongTVQA_plus_train.json` and `LongTVQA_plus_val.json` is a dictionary with the following fields:
| Key | Type | Description |
| --- | --- | --- |
| `qid` | int | Question ID (same as in TVQA+). |
| `q` | str | Question text. |
| `a0` ... `a4` | str | Five multiple-choice answers. |
| `answer` | str | Correct answer key (`"a0"``"a4"`). |
| `ts` | list | Refined timestamp annotation. For example, `[0, 5.4]` indicates the localized temporal span starts at 0s and ends at 5.4s. |
| `episode_name` | str | Episode ID (e.g. `s01e02`). |
| `occur_clip` | str | Video clip name. Format: `{show_name_abbr}_s{season}e{episode}_seg{segment}_clip_{clip}`. Episodes are typically divided into two segments separated by the opening theme. For **The Big Bang Theory**, `{show_name_abbr}` is omitted (e.g. `s05e02_seg02_clip_00`). |
| `bbox` | dict | Frame-level bounding box annotations sampled at 3 FPS. Keys are frame indices. Values are lists of bounding boxes with `img_id`, `top`, `left`, `width`, `height`, and `label`. |
---
### QA Sample
```json
{
"answer": "a1",
"qid": 134094,
"ts": [5.99, 11.98],
"a1": "Howard is talking to Raj and Leonard",
"a0": "Howard is talking to Bernadette",
"a3": "Howard is talking to Leonard and Penny",
"a2": "Howard is talking to Sheldon , and Raj",
"q": "Who is Howard talking to when he is in the lab room ?",
"episode_name": "s05e02",
"occur_clip": "s05e02_seg02_clip_00",
"a4": "Howard is talking to Penny and Bernadette",
"bbox": {
"14": [
{
"img_id": 14,
"top": 153,
"label": "Howard",
"width": 180,
"height": 207,
"left": 339
},
{
"img_id": 14,
"top": 6,
"label": "lab",
"width": 637,
"height": 354,
"left": 3
}
],
"20": [],
"26": [],
"32": [],
"38": []
}
}
```
---
## Subtitles JSON Format
Two subtitle files are provided to support different temporal granularities:
| File | Key | Type | Description |
| ------------------------------------------- | -------------- | ---- | ------------------------------------------------------------------------------------------------------------------- |
| `LongTVQA_plus_subtitle_clip_level.json` | `vid_name` | str | Clip-level subtitle text, with utterances separated by `<eos>`. |
| `LongTVQA_plus_subtitle_episode_level.json` | `episode_name` | str | Episode-level subtitle text, including clip markers such as `<seg01_clip_00>`, and utterances separated by `<eos>`. |
---
### Subtitles Sample
```json
{
"s09e14_seg02_clip_04": "Sheldon : That 's a risk I'm willing to take ! <eos> Amy : Well , this is so nice . <eos> ..."
}
```
---
## License
This dataset is released under the **MIT License**.
## 📝 Citation
If you find our work helpful, please cite:
```bibtex
@misc{liu2025longvideoagentmultiagentreasoninglong,
title={LongVideoAgent: Multi-Agent Reasoning with Long Videos},
author={Runtao Liu and Ziyi Liu and Jiaqi Tang and Yue Ma and Renjie Pi and Jipeng Zhang and Qifeng Chen},
year={2025},
eprint={2512.20618},
archivePrefix={arXiv},
primaryClass={cs.AI},
url={[https://arxiv.org/abs/2512.20618](https://arxiv.org/abs/2512.20618)},
}