# BBox DocVQA **Train Set** The BBox DocVQA Train Set is a large-scale dataset designed for training document visual question answering models with grounded supervision. Each QA instance is paired with one or more rendered PDF pages and pixel-level bounding boxes that mark the evidence required to answer the question. The dataset covers a broad distribution of document types, visual regions, and multi-page reasoning patterns. --- ## Repository layout The dataset is organized as follows: - **`BBox_DocVQA_Train.jsonl`** – newline-delimited JSON containing all training QA samples and metadata. - **`//*.png`** – rendered PDF pages grouped into eight arXiv subject categories (`cs`, `econ`, `eess`, `math`, `physics`, `q-bio`, `q-fin`, `stat`). - Page images follow the naming format: **`_.png`**, where `` corresponds to the original PDF’s 1-based page index. This directory layout mirrors the benchmark structure for seamless integration. --- ## Dataset statistics The BBox DocVQA Train Set contains: - **Total QA samples:** 30,780 - **Total pages:** 42,380 - **Total papers:** 3,671 ### Task type distribution | Task Type | Count | |----------|------:| | SPSBB | 11,668 (37.91%) | | SPMBB | 7,512 (24.41%) | | MPMBB | 11,600 (37.69%) | ### Region type distribution | Region Type | Count | |-------------|------:| | Text | 30,424 (60.98%) | | Image | 12,542 (25.14%) | | Table | 6,926 (13.88%) | - **Average bounding box area ratio:** 14.26% --- ## JSON lines schema Each entry in `BBox_DocVQA_Train.jsonl` follows the schema below: | Field | Type | Description | |-------|------|-------------| | `query` / `question` | string | Natural-language question (duplicate keys for compatibility). | | `answer` | string | Grounded short-form answer. | | `category` | string | One of the eight arXiv subject classes. | | `doc_name` | string | ArXiv identifier of the source paper. | | `evidence_page` | list[int] | Pages containing the evidence (1-based). | | `image_paths` / `images` | list[str] | Relative paths to one or two rendered PDF pages. | | `bbox` | list[list[list[int]]] | Bounding boxes for each referenced page, in pixel units. | | `subimg_tpye` | list[list[str]] | Region type per bounding box (`text`, `table`, or `image`). | --- ## Example ```json { "query": "What is the caption of Figure 3 on the referenced page?", "answer": "Comparison between the baseline and our method", "doc_name": "2301.12345", "category": "cs", "evidence_page": [4], "image_paths": ["cs/2301.12345/2301.12345_4.png"], "bbox": [ [[512, 1340, 1880, 1620]] ], "subimg_tpye": [["image"]] } ``` --- ## Quick start ```python import json from PIL import Image, ImageDraw with open("BBox_DocVQA_Train.jsonl") as f: sample = json.loads(f.readline()) for page_path, boxes in zip(sample["image_paths"], sample["bbox"]): img = Image.open(page_path).convert("RGB") draw = ImageDraw.Draw(img) for (xmin, ymin, xmax, ymax) in boxes: draw.rectangle((xmin, ymin, xmax, ymax), outline="red", width=5) img.show() ``` --- ## Notes and usage guidance - Page images are uncompressed PNG renders produced from arXiv PDFs; please observe arXiv’s terms of use for any redistribution. - Bounding boxes are provided in absolute pixel coordinates; normalize them by image width/height when required. - Duplicate key names (e.g., `query`/`question`, `image_paths`/`images`) are intentionally preserved for compatibility. - The train set provides large-scale grounded supervision across diverse document layouts and visual evidence types.