diff --git "a/2025/DELOC_ Document Element Localizer/layout.json" "b/2025/DELOC_ Document Element Localizer/layout.json" new file mode 100644--- /dev/null +++ "b/2025/DELOC_ Document Element Localizer/layout.json" @@ -0,0 +1,8589 @@ +{ + "pdf_info": [ + { + "para_blocks": [ + { + "bbox": [ + 176, + 76, + 418, + 92 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 176, + 76, + 418, + 92 + ], + "spans": [ + { + "bbox": [ + 176, + 76, + 418, + 92 + ], + "type": "text", + "content": "DELOC: Document Element Localizer" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "spans": [ + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "text", + "content": "Hammad Ayyubi" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "inline_equation", + "content": "^{1*}" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "text", + "content": ", Puneet Mathur" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "inline_equation", + "content": "^{2}" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "text", + "content": ", Md Mehrab Tanjim" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "inline_equation", + "content": "^{2}" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "text", + "content": ", Vlad I Morariu" + }, + { + "bbox": [ + 90, + 120, + 506, + 136 + ], + "type": "inline_equation", + "content": "^{2}" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 193, + 137, + 401, + 153 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 193, + 137, + 401, + 153 + ], + "spans": [ + { + "bbox": [ + 193, + 137, + 401, + 153 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 193, + 137, + 401, + 153 + ], + "type": "text", + "content": "Columbia University, " + }, + { + "bbox": [ + 193, + 137, + 401, + 153 + ], + "type": "inline_equation", + "content": "^{2}" + }, + { + "bbox": [ + 193, + 137, + 401, + 153 + ], + "type": "text", + "content": "Adobe Research," + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 213, + 154, + 380, + 166 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 213, + 154, + 380, + 166 + ], + "spans": [ + { + "bbox": [ + 213, + 154, + 380, + 166 + ], + "type": "text", + "content": "Correspondence: hayyubi@cs.columbia.edu" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 155, + 219, + 203, + 232 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 155, + 219, + 203, + 232 + ], + "spans": [ + { + "bbox": [ + 155, + 219, + 203, + 232 + ], + "type": "text", + "content": "Abstract" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 84, + 243, + 274, + 553 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 84, + 243, + 274, + 553 + ], + "spans": [ + { + "bbox": [ + 84, + 243, + 274, + 553 + ], + "type": "text", + "content": "Editing documents and PDFs using natural language instructions is desirable for many reasons - ease of use, increasing accessibility to non-technical users, and for creativity. To do this automatically, a system needs to first understand the user's intent and convert this to an executable plan or command, and then the system needs to identify or localize the elements that the user desires to edit. While there exist methods that can accomplish these tasks, a major bottleneck in these systems is the inability to ground the spatial edit location effectively. We address this gap through our proposed system, DELOC (Document Element LOCALizer). DELOC adapts the grounding capabilities of existing Multimodal Large Language Model (MLLM) from natural images to PDFs. This adaptation involves two novel contributions: 1) synthetically generating PDF-grounding instruction tuning data from partially annotated datasets; and 2) synthetic data cleaning via Code-NLI, an NLI-inspired process to clean data using generated Python code. The effectiveness of DELOC is apparent in the " + }, + { + "bbox": [ + 84, + 243, + 274, + 553 + ], + "type": "inline_equation", + "content": ">2\\mathrm{x}" + }, + { + "bbox": [ + 84, + 243, + 274, + 553 + ], + "type": "text", + "content": " zero-shot improvement it achieves over the next best MLLM, GPT-4o." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 68, + 564, + 155, + 576 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 564, + 155, + 576 + ], + "spans": [ + { + "bbox": [ + 68, + 564, + 155, + 576 + ], + "type": "text", + "content": "1 Introduction" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 585, + 290, + 665 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 585, + 290, + 665 + ], + "spans": [ + { + "bbox": [ + 67, + 585, + 290, + 665 + ], + "type": "text", + "content": "Editing documents and PDFs via natural language (Mathur et al., 2023; Suri et al., 2024) is an innovative and user-friendly advancement. This technology makes it easier for non-technical people to edit PDFs, speeds up the editing process, and facilitates document editing on a mobile device." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 67, + 666, + 290, + 747 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 666, + 290, + 747 + ], + "spans": [ + { + "bbox": [ + 67, + 666, + 290, + 747 + ], + "type": "text", + "content": "Typically, a two-stage process has been followed to edit PDFs from natural language requests. In the first stage, a model is required to predict the edit location (green highlight in Figure 1) spatially in the PDF. The edit location is a PDF element (paragraph, line, list, table, etc.). The second stage" + } + ] + } + ], + "index": 8 + }, + { + "type": "image", + "bbox": [ + 307, + 219, + 524, + 378 + ], + "blocks": [ + { + "bbox": [ + 307, + 219, + 524, + 378 + ], + "lines": [ + { + "bbox": [ + 307, + 219, + 524, + 378 + ], + "spans": [ + { + "bbox": [ + 307, + 219, + 524, + 378 + ], + "type": "image", + "image_path": "3471bc165d1c26d7258906fbdfb45df050adfad0e5090df4880608a459c0b1e1.jpg" + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 302, + 382, + 526, + 465 + ], + "lines": [ + { + "bbox": [ + 302, + 382, + 526, + 465 + ], + "spans": [ + { + "bbox": [ + 302, + 382, + 526, + 465 + ], + "type": "text", + "content": "Figure 1: Illustration of a typical PDF edit process. Models predict Edit Location and Edit Command, which are combined to effect the required edit. This work focuses on improving the bottleneck step - Edit Location. To this end, the proposed system, DELOC, generates large-scale (3M) synthetic data to instruction tune a Multimodal LLM." + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_caption" + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "spans": [ + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "type": "text", + "content": "involves predicting the edit command (replace in Figure 1). The predicted bounding box and edit command can then be simply combined to effect the desired change. Existing PDF editing systems achieve high accuracy " + }, + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "type": "inline_equation", + "content": "(>85\\%)" + }, + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "type": "text", + "content": " in predicting edit commands (Mathur et al., 2023) but struggle with bounding box precision " + }, + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "type": "inline_equation", + "content": "(<50\\%)" + }, + { + "bbox": [ + 302, + 483, + 525, + 605 + ], + "type": "text", + "content": " (Suri et al., 2024). As such, we focus on improving the spatial PDF element localization for a given edit request." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "spans": [ + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "type": "text", + "content": "State-of-the-art systems ground user queries in images (Wang et al., 2023; You et al., 2023), mobile UIs (You et al., 2024), and web UIs (Hong et al., 2023), but they do not transfer directly to PDFs. PDFs are more verbose, and edit requests follow a hierarchical structure requiring domain knowledge. For instance, a request like \"first list item of the third list in Section A.1\" demands an understanding of the composition: section " + }, + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "type": "inline_equation", + "content": "\\rightarrow" + }, + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "type": "text", + "content": " list " + }, + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "type": "inline_equation", + "content": "\\rightarrow" + }, + { + "bbox": [ + 302, + 609, + 525, + 730 + ], + "type": "text", + "content": " list item." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 735, + 525, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 735, + 525, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 735, + 525, + 775 + ], + "type": "text", + "content": "To address these challenges, we propose DELOC (Document Element LOCALizer), which adapts the strong grounding capabilities of Mul" + } + ] + } + ], + "index": 13 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 67, + 752, + 290, + 775 + ], + "type": "page_footnote", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 752, + 290, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 752, + 290, + 775 + ], + "type": "text", + "content": "*Work done during an internship at Adobe Research, Document Intelligence Lab (DIL)" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 282, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 282, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 282, + 780, + 312, + 791 + ], + "type": "text", + "content": "31127" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 91, + 795, + 502, + 818 + ], + "type": "footer", + "angle": 0, + "lines": [ + { + "bbox": [ + 91, + 795, + 502, + 818 + ], + "spans": [ + { + "bbox": [ + 91, + 795, + 502, + 818 + ], + "type": "text", + "content": "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 31127-31136 November 4-9, 2025 ©2025 Association for Computational Linguistics" + } + ] + } + ], + "index": 16 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 0 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 83, + 70, + 512, + 218 + ], + "blocks": [ + { + "bbox": [ + 83, + 70, + 512, + 218 + ], + "lines": [ + { + "bbox": [ + 83, + 70, + 512, + 218 + ], + "spans": [ + { + "bbox": [ + 83, + 70, + 512, + 218 + ], + "type": "image", + "image_path": "877ec0cf624bc8fc7f88120d472ecff5c21fda422b6ca29317517bcb9a1faebb.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 67, + 223, + 525, + 248 + ], + "lines": [ + { + "bbox": [ + 67, + 223, + 525, + 248 + ], + "spans": [ + { + "bbox": [ + 67, + 223, + 525, + 248 + ], + "type": "text", + "content": "Figure 2: Overview of DELOC. Document elements, their bounding box, and their hierarchical relations are first detected (DET) and formatted into a JSON structure (FJ). This JSON is fed to an LLM to generate synthetic data." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 261, + 290, + 342 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 261, + 290, + 342 + ], + "spans": [ + { + "bbox": [ + 67, + 261, + 290, + 342 + ], + "type": "text", + "content": "timodal Large Language Models (MLLMs) from natural images to PDFs. This adaptation involves instruction-tuning an MLLM specifically for PDFs. Our key contribution is the synthetic generation of high-quality instruction-tuning data, followed by an automated cleaning process called Code-NLI." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "spans": [ + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "type": "text", + "content": "We first represent PDF elements (paragraphs, lines, lists, tables, titles, sections) in a spatially aware hierarchical JSON format (Figure 2), capturing each element's bounding box and hierarchical relationships (e.g., paragraph " + }, + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "type": "inline_equation", + "content": "\\rightarrow" + }, + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "type": "text", + "content": " lines, list " + }, + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "type": "inline_equation", + "content": "\\rightarrow" + }, + { + "bbox": [ + 67, + 347, + 291, + 576 + ], + "type": "text", + "content": " list items). This information is sourced from partially annotated PDF Document Analysis datasets (e.g., PubLayNet (Zhong et al., 2019)), using existing annotations and heuristically generating missing ones. Next, we input this structured PDF representation into a Large Language Model (LLM) to generate synthetic user requests for PDF editing, along with system responses specifying edit locations via bounding boxes. The hierarchical structure enables the LLM to produce compositional requests that resemble real user queries, while the bounding box data ensures precise edit location predictions." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 581, + 290, + 689 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 581, + 290, + 689 + ], + "spans": [ + { + "bbox": [ + 67, + 581, + 290, + 689 + ], + "type": "text", + "content": "We further clean noisy generated samples using a proposed process called Code-NLI (Code-based Natural Language Inference). Code-NLI (Figure 3) treats the generated user edit request and system response as the hypothesis, and the PDF as the premise. To verify the hypothesis, it leverages an LLM to generate Python code, which is then executed to filter out noisy samples." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 694, + 291, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 694, + 291, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 694, + 291, + 775 + ], + "type": "text", + "content": "To demonstrate the effectiveness of DELOC, we evaluate it on DocEdit (Mathur et al., 2023). Our results demonstrate that DELOC outperforms all existing zero-shot models on PDFs, including proprietary MLLM GPT-4o by " + }, + { + "bbox": [ + 67, + 694, + 291, + 775 + ], + "type": "inline_equation", + "content": ">2\\mathrm{x}" + }, + { + "bbox": [ + 67, + 694, + 291, + 775 + ], + "type": "text", + "content": ". Our ablations demonstrate DELOC's performance improves with" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 261, + 526, + 288 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 261, + 526, + 288 + ], + "spans": [ + { + "bbox": [ + 302, + 261, + 526, + 288 + ], + "type": "text", + "content": "data scale and data cleaning via Code-NLI is crucial to its performance." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 299, + 396, + 312 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 299, + 396, + 312 + ], + "spans": [ + { + "bbox": [ + 302, + 299, + 396, + 312 + ], + "type": "text", + "content": "2 Related Work" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 301, + 322, + 526, + 456 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 301, + 322, + 526, + 456 + ], + "spans": [ + { + "bbox": [ + 301, + 322, + 526, + 456 + ], + "type": "text", + "content": "There has been a surge of interest in natural language based PDF edit localization. DocEdit (Mathur et al., 2023) trains a multimodal transformer that directly regresses the edit bounding box given the document as input. DocEdit-v2 (Suri et al., 2024) trains an encoder-decoder architecture that outputs a segmentation mask for the edit location. In training a model from scratch, these methods fail to utilize the excellent grounding capabilities of recent MLLMs." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 457, + 525, + 633 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 457, + 525, + 633 + ], + "spans": [ + { + "bbox": [ + 302, + 457, + 525, + 633 + ], + "type": "text", + "content": "A number of MLLMs (Chen et al., 2023; Zhang et al., 2023; Yuan et al., 2023; Lv et al., 2023; Lai et al., 2024; Ma et al., 2024) have been proposed that have shown impressive grounding abilities on natural images. Works like Ferret-UI (You et al., 2024), Cog-Agent (Hong et al., 2023), and MultiUI (Liu et al., 2024) have successfully adapted these MLLMs for mobile screens and web pages. We take inspiration from these works to adapt MLLMs for PDFs using large-scale synthetic data. In comparison, Agent-DocEdit (Wu et al., 2024) finetunes a grounding model, GLIP (Li* et al., 2022), only on a small training set." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 645, + 364, + 656 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 645, + 364, + 656 + ], + "spans": [ + { + "bbox": [ + 302, + 645, + 364, + 656 + ], + "type": "text", + "content": "3 Method" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 666, + 526, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 666, + 526, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 666, + 526, + 775 + ], + "type": "text", + "content": "Our method adapts the excellent query understanding and grounding capabilities of existing Multimodal Large Language Models (MLLMs) for grounding edit queries in PDFs. To this end, we: 1) generate diverse, compositional, and clean instruction-tuning data; and, 2) instruction-tune a MLLM on the generated data. We describe the synthetic data generation process below." + } + ] + } + ], + "index": 11 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31128" + } + ] + } + ], + "index": 12 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 1 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 71, + 74, + 182, + 211 + ], + "blocks": [ + { + "bbox": [ + 71, + 74, + 182, + 211 + ], + "lines": [ + { + "bbox": [ + 71, + 74, + 182, + 211 + ], + "spans": [ + { + "bbox": [ + 71, + 74, + 182, + 211 + ], + "type": "image", + "image_path": "ace1861ad173926c3c69e306ef5ebfeb2ba98dfbbb5ca3142b37385db11a6cf0.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 67, + 216, + 289, + 253 + ], + "lines": [ + { + "bbox": [ + 67, + 216, + 289, + 253 + ], + "spans": [ + { + "bbox": [ + 67, + 216, + 289, + 253 + ], + "type": "text", + "content": "Figure 3: Code-NLI based data cleaning. Generated data is treated as hypothesis, which is validated by LLM generated Python code, given PDF premise." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "type": "image", + "bbox": [ + 185, + 74, + 279, + 201 + ], + "blocks": [ + { + "bbox": [ + 185, + 74, + 279, + 201 + ], + "lines": [ + { + "bbox": [ + 185, + 74, + 279, + 201 + ], + "spans": [ + { + "bbox": [ + 185, + 74, + 279, + 201 + ], + "type": "image", + "image_path": "be6a08206f968efc123b8bfa78b055b3bcd2634e1051cb93482a7cd6ed97be57.jpg" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 210, + 204, + 258, + 212 + ], + "lines": [ + { + "bbox": [ + 210, + 204, + 258, + 212 + ], + "spans": [ + { + "bbox": [ + 210, + 204, + 258, + 212 + ], + "type": "text", + "content": "Generated Code" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_caption" + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 269, + 289, + 322 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 269, + 289, + 322 + ], + "spans": [ + { + "bbox": [ + 67, + 269, + 289, + 322 + ], + "type": "text", + "content": "Given a lack of large-scale instruction-tuning data for grounding edit queries in PDFs, we build an automatic synthetic data generation process. We discuss below the two stages of this process:" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 326, + 290, + 596 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 326, + 290, + 596 + ], + "spans": [ + { + "bbox": [ + 67, + 326, + 290, + 596 + ], + "type": "text", + "content": "LLM powered data generation. We present the PDF to an LLM and prompt it to generate edit requests along with the corresponding bounding box (see Figure 2). Prompt details in Appendix B. To ensure that the generated edit requests capture the compositional nature of real user queries, we structure PDFs in a hierarchical JSON format that explicitly defines the parent-child relationships of each element. Each element contains information about its type (e.g., paragraph, line, list), its bounding box within the PDF, and its subelements (e.g., a line is a subelement of a paragraph). Additionally, we organize elements sequentially within the same hierarchy based on their y-coordinates, enabling the LLM to generate edit requests that reference these elements numerically (e.g., \"second list,\" \"last line\"). Moreover, the inclusion of a bounding box for each element allows the LLM to generate precise edit location in the system response for the corresponding edit query." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 66, + 599, + 291, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 66, + 599, + 291, + 773 + ], + "spans": [ + { + "bbox": [ + 66, + 599, + 291, + 773 + ], + "type": "text", + "content": "To obtain this hierarchical PDF representation, we require annotations for each PDF element, including its type, bounding box, and hierarchical subelements. However, there does not exist a single dataset that contains all this information for all element types. As such, we use different datasets - PubLayNet and DocLayNet (Pfitzmann et al., 2022) - each containing a subset of these elements. Table 1 shows a non-exhaustive list of these elements and the corresponding dataset containing them. If an element's hierarchical subelements' annotations do not exist natively in the dataset, we use heuristics to create them. These heuristics are rela" + } + ] + } + ], + "index": 6 + }, + { + "type": "table", + "bbox": [ + 305, + 69, + 523, + 131 + ], + "blocks": [ + { + "bbox": [ + 305, + 69, + 523, + 131 + ], + "lines": [ + { + "bbox": [ + 305, + 69, + 523, + 131 + ], + "spans": [ + { + "bbox": [ + 305, + 69, + 523, + 131 + ], + "type": "table", + "html": "
TitleParagraphListTablesFigures
DatasetsLinesList-items
Lines
PubLayNet
DocLayNet
", + "image_path": "b6e7adcf3577c9029b3bd601ae99b0fd15831e89da74663a6c72aabc9558d28d.jpg" + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "table_body" + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 134, + 525, + 170 + ], + "lines": [ + { + "bbox": [ + 302, + 134, + 525, + 170 + ], + "spans": [ + { + "bbox": [ + 302, + 134, + 525, + 170 + ], + "type": "text", + "content": "Table 1: Datasets and their PDF elements - either natively (✓) present or heuristically created (♠). Multiple datasets combine to give extensive elements coverage." + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 302, + 184, + 525, + 306 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 184, + 525, + 306 + ], + "spans": [ + { + "bbox": [ + 302, + 184, + 525, + 306 + ], + "type": "text", + "content": "tively simple - comparing y-coordinates of words to create lines within paragraphs, comparing x-coordinates of lines to create list-items within lists, and so on. The granularity scope goes as low as words. This means we have annotations for words, equations, formulae, footnotes, page numbers, etc. Essentially, we tried to cover as many PDF elements as possible with the available datasets and heuristics. More details in Appendix A." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 306, + 525, + 441 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 306, + 525, + 441 + ], + "spans": [ + { + "bbox": [ + 302, + 306, + 525, + 441 + ], + "type": "text", + "content": "Data cleaning with Code-NLI. Since the LLM generates data automatically, it can be noisy—for instance, an edit request might reference the second list item while the bounding box corresponds to the third. To filter out such inconsistencies, we draw inspiration from Natural Language Inference (NLI) (Bowman et al., 2015), which predicts whether a hypothesis aligns with a given premise. Here, the premise is the PDF, and the hypothesis is the generated user-system conversation." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 443, + 525, + 577 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 443, + 525, + 577 + ], + "spans": [ + { + "bbox": [ + 302, + 443, + 525, + 577 + ], + "type": "text", + "content": "To verify these conversations, we generate substatements that must hold true. For example, to confirm that a bounding box corresponds to the second list item, a counting statement must evaluate its position as 2 (Figure 3). These sub-statements are generated by an LLM as Pythonic function calls, which are then executed using our implemented APIs for verification. Pythonic statements make verification algorithmic and automatic. API details and prompts in Appendix C." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 578, + 524, + 632 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 578, + 524, + 632 + ], + "spans": [ + { + "bbox": [ + 302, + 578, + 524, + 632 + ], + "type": "text", + "content": "Once we have clean synthetic instruction-tuning data, we finetune a MLLM. The input is an image of the PDF and a user-edit request. The output is a bounding box localizing the edit element." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 644, + 389, + 658 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 644, + 389, + 658 + ], + "spans": [ + { + "bbox": [ + 302, + 644, + 389, + 658 + ], + "type": "text", + "content": "4 Experiments" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "type": "text", + "content": "To balance quality and the cost of long context, we use LLaMA-3.1 70B (Grattafori and Team, 2024) for generating synthetic data. It is run on 40 Nvidia A100-80 GiB GPUs for 2 days. Next, GPT-4o (OpenAI, 2024) is used in Code-NLI cleaning as small context length allows us to optimize for quality. This results in " + }, + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "type": "inline_equation", + "content": "\\sim 3\\mathrm{M}" + }, + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "type": "text", + "content": " samples for instruction tuning. We keep " + }, + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "type": "inline_equation", + "content": "95\\%" + }, + { + "bbox": [ + 302, + 666, + 525, + 775 + ], + "type": "text", + "content": " samples for training and the rest" + } + ] + } + ], + "index": 14 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31129" + } + ] + } + ], + "index": 15 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 2 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 70, + 69, + 289, + 211 + ], + "blocks": [ + { + "bbox": [ + 70, + 69, + 289, + 211 + ], + "lines": [ + { + "bbox": [ + 70, + 69, + 289, + 211 + ], + "spans": [ + { + "bbox": [ + 70, + 69, + 289, + 211 + ], + "type": "table", + "html": "
ModelPDFDesign
A@0.5A@0.30A@0.25A@0.5A@0.30A@0.25
Zero-Shot
GPT-4o5.62414.25817.88528.96146.76652.582
CogAgent4.169.4211.13---
Qwen2.5-VL2.906.978.72---
Ferret-UI0.00.220.28---
Phi-3-V0.4502.7274.18910.42325.26431.316
DELOC14.70330.39135.19832.42151.48957.445
Finetuned
DocEdit36.500--34.340--
DocEdit-v248.690-----
Phi-3-V26.56748.52454.37134.52357.18362.272
DELOC49.62067.02369.97557.01271.56674.976
", + "image_path": "78035f76c3c0889048a9644ed344a3f10fd2a246da4a03660acc3e10ed03a3a3.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "type": "table", + "bbox": [ + 114, + 282, + 245, + 322 + ], + "blocks": [ + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "lines": [ + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "spans": [ + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "type": "text", + "content": "Table 2: Comparison of DELOC with state-of-the-art on DocEdit edit location (bbox) prediction. DELOC outperforms both existing Multimodal LLMs in zeroshot setting and specialized models in finetuned setting. " + }, + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "type": "inline_equation", + "content": "A@X" + }, + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "type": "text", + "content": " denotes Accuracy at IoU of " + }, + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "type": "inline_equation", + "content": "X" + }, + { + "bbox": [ + 67, + 213, + 291, + 274 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 114, + 282, + 245, + 322 + ], + "lines": [ + { + "bbox": [ + 114, + 282, + 245, + 322 + ], + "spans": [ + { + "bbox": [ + 114, + 282, + 245, + 322 + ], + "type": "table", + "html": "
ModelA@0.5
DELOC6.297
- w/o CODE-NLI Filtering5.848
", + "image_path": "f641877bf18a122184a0e00ce187125fa3ab840f24ca1622018db45c10563752.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "table_body" + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 326, + 291, + 350 + ], + "lines": [ + { + "bbox": [ + 67, + 326, + 291, + 350 + ], + "spans": [ + { + "bbox": [ + 67, + 326, + 291, + 350 + ], + "type": "text", + "content": "Table 3: Ablation of Code-NLI. The filtering step improves DELOC performance." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 67, + 364, + 290, + 379 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 364, + 290, + 379 + ], + "spans": [ + { + "bbox": [ + 67, + 364, + 290, + 379 + ], + "type": "text", + "content": "for validation. More data statistics in Appendix D." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 380, + 290, + 488 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 380, + 290, + 488 + ], + "spans": [ + { + "bbox": [ + 67, + 380, + 290, + 488 + ], + "type": "text", + "content": "For instruction-tuning Phi-3V (Abdin and Team, 2024) is used as it's small and thus easy to train. We use a learning rate of 5e-5, weight decay of 0.12, AdamW optimizer, and a linear scheduler with a warmup. It takes approximately 10 hours to train the model on 128 Nvidia A100-80 GiB GPUs. More architecture and training details are in Appendix E and Appendix F." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 489, + 291, + 731 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 489, + 291, + 731 + ], + "spans": [ + { + "bbox": [ + 67, + 489, + 291, + 731 + ], + "type": "text", + "content": "We evaluate our proposed approach on the test set of the DocEdit dataset (Mathur et al., 2023). The test is split into two subsets: PDFs, which are more verbose, and Designs, which are graphical. We use the originally proposed metric of calculating accuracy by evaluating if the Intersection Over Union (IoU) between predicted bbox and groundtruth bbox is greater than a threshold. For baselines, we select state-of-the-art MLLMs that understand text well from images. Consequently, we select GPT-4o, Phi3-V, CogAgent, Ferret-UI, Qwen2.5-VL (Bai et al., 2025) and Fuyu (Bavishi et al., 2023). We also consider as baselines specialized models - DocEdit v1 and v2 - that are finetuned on DocEdit. We compare DELOC against these baselines in two settings: zero-shot - where DELOC is not finetuned on the DocEdit train set, and finetuned - where it is." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 735, + 291, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 735, + 291, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 735, + 291, + 775 + ], + "type": "text", + "content": "Our main results are summarized in Table 2. We make the following observations: 1) DELOC outperforms both open-source and proprietary closed" + } + ] + } + ], + "index": 7 + }, + { + "type": "image", + "bbox": [ + 336, + 74, + 495, + 174 + ], + "blocks": [ + { + "bbox": [ + 336, + 74, + 495, + 174 + ], + "lines": [ + { + "bbox": [ + 336, + 74, + 495, + 174 + ], + "spans": [ + { + "bbox": [ + 336, + 74, + 495, + 174 + ], + "type": "image", + "image_path": "346a7c830ca3cd72211ea5d8dd72aab6483a319921ba7abd1e077baed3bca2b5.jpg" + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 302, + 179, + 524, + 202 + ], + "lines": [ + { + "bbox": [ + 302, + 179, + 524, + 202 + ], + "spans": [ + { + "bbox": [ + 302, + 179, + 524, + 202 + ], + "type": "text", + "content": "Figure 4: Datasize Ablation: DELOC's performance improves with data." + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_caption" + } + ], + "index": 8 + }, + { + "type": "image", + "bbox": [ + 311, + 216, + 523, + 386 + ], + "blocks": [ + { + "bbox": [ + 311, + 216, + 523, + 386 + ], + "lines": [ + { + "bbox": [ + 311, + 216, + 523, + 386 + ], + "spans": [ + { + "bbox": [ + 311, + 216, + 523, + 386 + ], + "type": "image", + "image_path": "049356e43ffae248e45e761dc8d937f584b751c1303a1426d4329383dc78a6d1.jpg" + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 302, + 391, + 524, + 403 + ], + "lines": [ + { + "bbox": [ + 302, + 391, + 524, + 403 + ], + "spans": [ + { + "bbox": [ + 302, + 391, + 524, + 403 + ], + "type": "text", + "content": "Figure 5: Qualitative sample prediction from DELOC." + } + ] + } + ], + "index": 11, + "angle": 0, + "type": "image_caption" + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "spans": [ + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "content": "source MLLM by " + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "inline_equation", + "content": "\\sim 3\\mathrm{x}" + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "content": " on the PDF subset and " + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "inline_equation", + "content": "\\sim 12\\%" + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "content": " on the Design subset. 2) DELOC outperforms all specialized finetuned models on both subsets. 3) DELOC improves over its backbone MLLM, Phi-3-V by " + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "inline_equation", + "content": "28\\mathrm{x}" + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "content": " on PDFs and " + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "inline_equation", + "content": "3\\mathrm{x}" + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "content": " on Designs, underscoring the significance of our synthetic instruction tuning. 4) The best existing zero-shot MLLM performance is at best " + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "inline_equation", + "content": "5\\%" + }, + { + "bbox": [ + 302, + 419, + 525, + 661 + ], + "type": "text", + "content": ", indicating grounding in PDF is challenging. 5) All MLLM perform much better on the Design subset. This is understandable given all MLLM have been pretrained on natural images, which are closer in domain to Designs than to verbose PDFs. Notably, we found Fuyu to be quite bad; hence we do not include it in Table 2. We also don't compare against Agent-DocEdit as it includes an additional post-processing step that assumes availability of ground-truthbbox for all elements in a PDF." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "spans": [ + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "text", + "content": "Our ablation on Code-NLI in Table 3 indicates the importance of data cleaning to our approach. We also manually evaluate the quality of Code-NLI on 25 samples ourselves. Our findings are: 1) Precision is " + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "inline_equation", + "content": "84\\%" + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "text", + "content": ": This implies that most of the data that Code-NLI predicts as correct is actually correct; only " + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "inline_equation", + "content": "16\\%" + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "text", + "content": " incorrect data passes through the Code-NLI filter. 2) Recall is " + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "inline_equation", + "content": "63.64\\%" + }, + { + "bbox": [ + 302, + 666, + 525, + 773 + ], + "type": "text", + "content": ": This" + } + ] + } + ], + "index": 13 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31130" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 3 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 290, + 152 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 290, + 152 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 290, + 152 + ], + "type": "text", + "content": "implies it rejects a decent amount of good data " + }, + { + "bbox": [ + 67, + 71, + 290, + 152 + ], + "type": "inline_equation", + "content": "(36\\%)" + }, + { + "bbox": [ + 67, + 71, + 290, + 152 + ], + "type": "text", + "content": ". This loss is acceptable for our use case since we generate a very large amount of synthetic data. The important consideration for us is that incorrect data should not be flagged as correct – which Code-NLI does (84% precision)." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 153, + 291, + 248 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 153, + 291, + 248 + ], + "spans": [ + { + "bbox": [ + 67, + 153, + 291, + 248 + ], + "type": "text", + "content": "Figure 4 demonstrates that performance of DELOC improves with data size. We also provide a qualitative example in Figure 5 that shows 1) DELOC acquires effective capability of verbose PDF grounding. 2) Ambiguity in the expected response can lead to divergence from ground-truth. More examples in Appendix G." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 256, + 147, + 269 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 256, + 147, + 269 + ], + "spans": [ + { + "bbox": [ + 67, + 256, + 147, + 269 + ], + "type": "text", + "content": "5 Conclusion" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 276, + 291, + 384 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 276, + 291, + 384 + ], + "spans": [ + { + "bbox": [ + 67, + 276, + 291, + 384 + ], + "type": "text", + "content": "In this work, we propose DELOC, a system to ground PDF edit request spatially. DELOC leverages spatial and hierarchical structure of PDFs to generate 3M synthetic instruction tuning data that successfully adapts existing MLLM for PDF grounding. Our results demonstrate that DELOC outperforms both existing MLLMs, including GPT-40, and specialized models on DocEdit." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 394, + 131, + 406 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 394, + 131, + 406 + ], + "spans": [ + { + "bbox": [ + 67, + 394, + 131, + 406 + ], + "type": "text", + "content": "Limitations" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 414, + 291, + 671 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 414, + 291, + 671 + ], + "spans": [ + { + "bbox": [ + 67, + 414, + 291, + 671 + ], + "type": "text", + "content": "We attempted to cover a comprehensive list of PDF elements in our synthetically generated data. This coverage is mostly dependent on the element annotation present in PDF analysis datasets we began with to create our hierarchical representation. As such, there could be some PDF elements which are not covered. We leave for future work the addition of more PDF analysis datasets to increase element coverage. Furthermore, this method (like most grounding MLLM approaches) generates bounding boxes using auto-regressive prediction. The loss used assigns equal weight to all predictions not exactly same as ground-truthbbox, irrespective of their proximity to the ground-truthbbox. Ideally, the loss should be higher for a predictedbbox that is farther from the ground-truth than a predicted box that is closer to ground-truth. Reinforcement learning based preference optimization could be a way to address this issue in future works." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 693, + 127, + 705 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 693, + 127, + 705 + ], + "spans": [ + { + "bbox": [ + 69, + 693, + 127, + 705 + ], + "type": "text", + "content": "References" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 711, + 291, + 775 + ], + "type": "list", + "angle": 0, + "index": 9, + "blocks": [ + { + "bbox": [ + 69, + 711, + 290, + 745 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 711, + 290, + 745 + ], + "spans": [ + { + "bbox": [ + 69, + 711, + 290, + 745 + ], + "type": "text", + "content": "Marah Abdin and The Phi-3 Team. 2024. Phi-3 technical report: A highly capable language model locally on your phone. Preprint, arXiv:2404.14219." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 751, + 291, + 775 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 751, + 291, + 775 + ], + "spans": [ + { + "bbox": [ + 69, + 751, + 291, + 775 + ], + "type": "text", + "content": "Shuai Bai, Keqin Chen, Xuejing Liu, Jialin Wang, Wenbin Ge, Sibo Song, Kai Dang, Peng Wang," + } + ] + } + ], + "index": 8 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 526, + 775 + ], + "type": "list", + "angle": 0, + "index": 21, + "blocks": [ + { + "bbox": [ + 314, + 72, + 525, + 148 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 72, + 525, + 148 + ], + "spans": [ + { + "bbox": [ + 314, + 72, + 525, + 148 + ], + "type": "text", + "content": "Shijie Wang, Jun Tang, Humen Zhong, Yanzhi Zhu, Mingkun Yang, Zhaohai Li, Jianqiang Wan, Pengfei Wang, Wei Ding, Zheren Fu, Yiheng Xu, Jiabo Ye, Xi Zhang, Tianbao Xie, Zesen Cheng, Hang Zhang, Zhibo Yang, Haiyang Xu, and Junyang Lin. 2025. Qwen2.5-vl technical report. Preprint, arXiv:2502.13923." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 304, + 158, + 525, + 201 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 158, + 525, + 201 + ], + "spans": [ + { + "bbox": [ + 304, + 158, + 525, + 201 + ], + "type": "text", + "content": "Rohan Bavishi, Erich Elsen, Curtis Hawthorne, Maxwell Nye, Augustus Odena, Arushi Somani, and Săgnak Taşürlar. 2023. Introducing our multimodal models." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 304, + 210, + 526, + 287 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 210, + 526, + 287 + ], + "spans": [ + { + "bbox": [ + 304, + 210, + 526, + 287 + ], + "type": "text", + "content": "Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 632-642, Lisbon, Portugal. Association for Computational Linguistics." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 296, + 525, + 340 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 296, + 525, + 340 + ], + "spans": [ + { + "bbox": [ + 304, + 296, + 525, + 340 + ], + "type": "text", + "content": "Keqin Chen, Zhao Zhang, Weili Zeng, Richong Zhang, Feng Zhu, and Rui Zhao. 2023. Shikra: Unleashing multimodal llm's referential dialogue magic. arXiv preprint arXiv:2306.15195." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 348, + 525, + 457 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 348, + 525, + 457 + ], + "spans": [ + { + "bbox": [ + 304, + 348, + 525, + 457 + ], + "type": "text", + "content": "Xiaoyi Dong, Pan Zhang, Yuhang Zang, Yuhang Cao, Bin Wang, Linke Ouyang, Songyang Zhang, Haodong Duan, Wenwei Zhang, Yining Li, Hang Yan, Yang Gao, Zhe Chen, Xinyue Zhang, Wei Li, Jingwen Li, Wenhai Wang, Kai Chen, Conghui He, Xingcheng Zhang, Jifeng Dai, Yu Qiao, Dahua Lin, and Jiaqi Wang. 2024. Internlm-xcomposer2-4khd: A pioneering large vision-language model handling resolutions from 336 pixels to 4k hd. Preprint, arXiv:2404.06512." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 466, + 525, + 489 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 466, + 525, + 489 + ], + "spans": [ + { + "bbox": [ + 304, + 466, + 525, + 489 + ], + "type": "text", + "content": "Aaron Grattafori and The LLaMA Team. 2024. The llama 3 herd of models. Preprint, arXiv:2407.21783." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 497, + 525, + 552 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 497, + 525, + 552 + ], + "spans": [ + { + "bbox": [ + 304, + 497, + 525, + 552 + ], + "type": "text", + "content": "Wenyi Hong, Weihan Wang, Qingsong Lv, Jiazheng Xu, Wenmeng Yu, Junhui Ji, Yan Wang, Zihan Wang, Yuxiao Dong, Ming Ding, and Jie Tang. 2023. Co-gagent: A visual language model for gui agents. Preprint, arXiv:2312.08914." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 560, + 525, + 605 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 560, + 525, + 605 + ], + "spans": [ + { + "bbox": [ + 304, + 560, + 525, + 605 + ], + "type": "text", + "content": "Xin Lai, Zhuotao Tian, Yukang Chen, Yanwei Li, Yuhui Yuan, Shu Liu, and Jiaya Jia. 2024. Lisa: Reasoning segmentation via large language model. Preprint, arXiv:2308.00692." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 613, + 525, + 669 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 613, + 525, + 669 + ], + "spans": [ + { + "bbox": [ + 304, + 613, + 525, + 669 + ], + "type": "text", + "content": "Lianian Harold Li*, Pengchuan Zhang*, Haotian Zhang*, Jianwei Yang, Chunyuan Li, Yiwu Zhong, Lijuan Wang, Lu Yuan, Lei Zhang, Jenq-Neng Hwang, Kai-Wei Chang, and Jianfeng Gao. 2022. Grounded language-image pre-training. In CVPR." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 677, + 525, + 731 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 677, + 525, + 731 + ], + "spans": [ + { + "bbox": [ + 304, + 677, + 525, + 731 + ], + "type": "text", + "content": "Junpeng Liu, Tianyue Ou, Yifan Song, Yuxiao Qu, Wai Lam, Chenyan Xiong, Wenhu Chen, Graham Neubig, and Xiang Yue. 2024. Harnessing webpage uis for text-rich visual understanding. Preprint, arXiv:2410.13824." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 740, + 525, + 775 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 740, + 525, + 775 + ], + "spans": [ + { + "bbox": [ + 304, + 740, + 525, + 775 + ], + "type": "text", + "content": "Tengchao Lv, Yupan Huang, Jingye Chen, Lei Cui, Shuming Ma, Yaoyao Chang, Shaohan Huang, Wenhui Wang, Li Dong, Weiyao Luo, et al. 2023." + } + ] + } + ], + "index": 20 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 311, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 311, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 311, + 791 + ], + "type": "text", + "content": "31131" + } + ] + } + ], + "index": 22 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 4 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 289, + 772 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 80, + 72, + 289, + 95 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 72, + 289, + 95 + ], + "spans": [ + { + "bbox": [ + 80, + 72, + 289, + 95 + ], + "type": "text", + "content": "Kosmos-2.5: A multimodal literate model. arXiv preprint arXiv:2309.11419." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 106, + 289, + 150 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 106, + 289, + 150 + ], + "spans": [ + { + "bbox": [ + 69, + 106, + 289, + 150 + ], + "type": "text", + "content": "Chuofan Ma, Yi Jiang, Jiannan Wu, Zehuan Yuan, and Xiaojuan Qi. 2024. Groma: Localized visual tokenization for grounding multimodal large language models. arXiv preprint arXiv:2404.13013." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 162, + 289, + 217 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 162, + 289, + 217 + ], + "spans": [ + { + "bbox": [ + 69, + 162, + 289, + 217 + ], + "type": "text", + "content": "Puneet Mathur, Rajiv Jain, Jiumiang Gu, Franck Dernoncourt, Dinesh Manocha, and Vlad Morariu. 2023. Docedit: Language-guided document editing. In Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI)." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 229, + 289, + 251 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 229, + 289, + 251 + ], + "spans": [ + { + "bbox": [ + 69, + 229, + 289, + 251 + ], + "type": "text", + "content": "OpenAI. 2024. Gpt-4o system card. Preprint, arXiv:2410.21276." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 263, + 289, + 329 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 263, + 289, + 329 + ], + "spans": [ + { + "bbox": [ + 69, + 263, + 289, + 329 + ], + "type": "text", + "content": "Birgit Pfitzmann, Christoph Auer, Michele Dolfi, Ahmed S. Nassar, and Peter Staar. 2022. Doclaynet: A large human-annotated dataset for documentlayout segmentation. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, page 3743-3751. ACM." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 340, + 289, + 407 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 340, + 289, + 407 + ], + "spans": [ + { + "bbox": [ + 69, + 340, + 289, + 407 + ], + "type": "text", + "content": "Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. 2021. Learning transferable visual models from natural language supervision. Preprint, arXiv:2103.00020." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 417, + 289, + 507 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 417, + 289, + 507 + ], + "spans": [ + { + "bbox": [ + 69, + 417, + 289, + 507 + ], + "type": "text", + "content": "Manan Suri, Puneet Mathur, Franck Dernoncourt, Rajiv Jain, Vlad I Morariu, Ramit Sawhney, Preslav Nakov, and Dinesh Manocha. 2024. DocEdit-v2: Document structure editing via multimodal LLM grounding. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 15485-15505, Miami, Florida, USA. Association for Computational Linguistics." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 518, + 289, + 584 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 518, + 289, + 584 + ], + "spans": [ + { + "bbox": [ + 69, + 518, + 289, + 584 + ], + "type": "text", + "content": "Weihan Wang, Qingsong Lv, Wenmeng Yu, Wenyi Hong, Ji Qi, Yan Wang, Junhui Ji, Zhuoyi Yang, Lei Zhao, Xixuan Song, Jiazheng Xu, Bin Xu, Juanzi Li, Yuxiao Dong, Ming Ding, and Jie Tang. 2023. Cogvlm: Visual expert for pretrained language models. Preprint, arXiv:2311.03079." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 596, + 289, + 640 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 596, + 289, + 640 + ], + "spans": [ + { + "bbox": [ + 69, + 596, + 289, + 640 + ], + "type": "text", + "content": "Te-Lin Wu, Rajiv Jain, Yufan Zhou, Puneet Mathur, and Vlad I Morariu. 2024. Agent-docedit: Language-instructed LLM agent for content-rich document editing. In First Conference on Language Modeling." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 651, + 289, + 707 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 651, + 289, + 707 + ], + "spans": [ + { + "bbox": [ + 69, + 651, + 289, + 707 + ], + "type": "text", + "content": "Haoxuan You, Haotian Zhang, Zhe Gan, Xianzhi Du, Bowen Zhang, Zirui Wang, Liangliang Cao, Shih-Fu Chang, and Yinfei Yang. 2023. Ferret: Refer and ground anything anywhere at any granularity. Preprint, arXiv:2310.07704." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 719, + 289, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 719, + 289, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 719, + 289, + 772 + ], + "type": "text", + "content": "Keen You, Haotian Zhang, Eldon Schoop, Floris Weers, Amanda Swearngin, Jeffrey Nichols, Yinfei Yang, and Zhe Gan. 2024. Ferret-ui: Grounded mobile ui understanding with multimodal llms. Preprint, arXiv:2404.05719." + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 305, + 72, + 524, + 200 + ], + "type": "list", + "angle": 0, + "index": 15, + "blocks": [ + { + "bbox": [ + 305, + 72, + 524, + 116 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 305, + 72, + 524, + 116 + ], + "spans": [ + { + "bbox": [ + 305, + 72, + 524, + 116 + ], + "type": "text", + "content": "Yuqian Yuan, Wentong Li, Jian Liu, Dongqi Tang, Xinjie Luo, Chi Qin, Lei Zhang, and Jianke Zhu. 2023. Osprey: Pixel understanding with visual instruction tuning. Preprint, arXiv:2312.10032." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 305, + 125, + 524, + 158 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 305, + 125, + 524, + 158 + ], + "spans": [ + { + "bbox": [ + 305, + 125, + 524, + 158 + ], + "type": "text", + "content": "Ao Zhang, Yuan Yao, Wei Ji, Zhiyuan Liu, and Tat-Seng Chua. 2023. Next-chat: An Imm for chat, detection and segmentation. Preprint, arXiv:2311.04498." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 305, + 167, + 524, + 200 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 305, + 167, + 524, + 200 + ], + "spans": [ + { + "bbox": [ + 305, + 167, + 524, + 200 + ], + "type": "text", + "content": "Xu Zhong, Jianbin Tang, and Antonio Jimeno Yepes. 2019. Publaynet: largest dataset ever for document layout analysis. Preprint, arXiv:1908.07836." + } + ] + } + ], + "index": 14 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 781, + 312, + 791 + ], + "type": "text", + "content": "31132" + } + ] + } + ], + "index": 16 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 5 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 290, + 97 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 290, + 97 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 290, + 97 + ], + "type": "text", + "content": "We provide additional details here for further clarification." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 80, + 108, + 291, + 282 + ], + "type": "list", + "angle": 0, + "index": 8, + "blocks": [ + { + "bbox": [ + 80, + 108, + 281, + 121 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 108, + 281, + 121 + ], + "spans": [ + { + "bbox": [ + 80, + 108, + 281, + 121 + ], + "type": "text", + "content": "- Element Creation Heuristics (Appendix A)" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 80, + 130, + 290, + 157 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 130, + 290, + 157 + ], + "spans": [ + { + "bbox": [ + 80, + 130, + 290, + 157 + ], + "type": "text", + "content": "- Synthetic Data Generation Prompt (Appendix B)" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 80, + 165, + 290, + 179 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 165, + 290, + 179 + ], + "spans": [ + { + "bbox": [ + 80, + 165, + 290, + 179 + ], + "type": "text", + "content": "Code-NLI Prompt and Samples (Appendix C)" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 80, + 188, + 220, + 201 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 188, + 220, + 201 + ], + "spans": [ + { + "bbox": [ + 80, + 188, + 220, + 201 + ], + "type": "text", + "content": "Data Statistics (Appendix D)" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 80, + 210, + 241, + 223 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 210, + 241, + 223 + ], + "spans": [ + { + "bbox": [ + 80, + 210, + 241, + 223 + ], + "type": "text", + "content": "- Model Architecture (Appendix E)" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 80, + 232, + 275, + 246 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 232, + 275, + 246 + ], + "spans": [ + { + "bbox": [ + 80, + 232, + 275, + 246 + ], + "type": "text", + "content": "Additional Training Details (Appendix F)" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 80, + 254, + 291, + 282 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 254, + 291, + 282 + ], + "spans": [ + { + "bbox": [ + 80, + 254, + 291, + 282 + ], + "type": "text", + "content": "Additional Qualitative Examples (Appendix G)" + } + ] + } + ], + "index": 7 + } + ], + "sub_type": "text" + }, + { + "type": "image", + "bbox": [ + 73, + 303, + 286, + 593 + ], + "blocks": [ + { + "bbox": [ + 73, + 303, + 286, + 593 + ], + "lines": [ + { + "bbox": [ + 73, + 303, + 286, + 593 + ], + "spans": [ + { + "bbox": [ + 73, + 303, + 286, + 593 + ], + "type": "image", + "image_path": "46a94597db36ae44aab255a15c85781bcea974a3a96ee472a3a25ffc18c9b672.jpg" + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 67, + 598, + 289, + 622 + ], + "lines": [ + { + "bbox": [ + 67, + 598, + 289, + 622 + ], + "spans": [ + { + "bbox": [ + 67, + 598, + 289, + 622 + ], + "type": "text", + "content": "Figure 6: Sample output from using our heuristic to create lines and list-items." + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_caption" + } + ], + "index": 9 + }, + { + "bbox": [ + 68, + 645, + 238, + 659 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 645, + 238, + 659 + ], + "spans": [ + { + "bbox": [ + 68, + 645, + 238, + 659 + ], + "type": "text", + "content": "A Element Creation Heuristics" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 67, + 666, + 291, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 666, + 291, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 666, + 291, + 775 + ], + "type": "text", + "content": "We require annotations for elements in PDFs to create the spatially-aware hierarchical representation. While an element annotation may be present in a dataset, its subelement or parent element may not have the requisite annotation available. We use heuristics to create annotations for such elements. To create lines within a paragraph, we take the bounding box of every word in the paragraph. All" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 71, + 525, + 179 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 525, + 179 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 525, + 179 + ], + "type": "text", + "content": "the words with same y-coordinates are considered to be in the same line. Similarly, to create lists from lines, we compare the x-coordinates of the beginning of the lines. Outliers are considered to be the beginning of a new list item. We illustrate the output from such heuristics to create lines and list-items in Figure 6. While these are simple, they work quite well." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 302, + 190, + 505, + 204 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 190, + 505, + 204 + ], + "spans": [ + { + "bbox": [ + 302, + 190, + 505, + 204 + ], + "type": "text", + "content": "B Synthetic Data Generation Prompt" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 302, + 211, + 526, + 332 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 211, + 526, + 332 + ], + "spans": [ + { + "bbox": [ + 302, + 211, + 526, + 332 + ], + "type": "text", + "content": "Once we have a hierarchical representation of the PDF elements along with their corresponding bounding box annotations, we feed them into an LLM to generate the instruction tuning data. Apart from the PDF elements, we input a system prompt containing the guidelines for generating an instruction tuning set and a one-shot example of how the user system conversation should be formatted. The process is illustrated in Figure 7." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 302, + 333, + 526, + 453 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 333, + 526, + 453 + ], + "spans": [ + { + "bbox": [ + 302, + 333, + 526, + 453 + ], + "type": "text", + "content": "The system prompt contains guidelines describing the task, input, and expected output. It contains instructions to generate diverse and compositional edit requests, to only utilize the information in the given hierarchical PDF representation, to not hallucinate, and so on. In addition, we provide a one-shot example to the LLM to further illustrate our requirements and guide it to follow expected format." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 314, + 455, + 483, + 468 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 455, + 483, + 468 + ], + "spans": [ + { + "bbox": [ + 314, + 455, + 483, + 468 + ], + "type": "text", + "content": "We provide the system prompt below:" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "spans": [ + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": "You are an AI visual assistance that can analyze PDFs. You will receive information describing a pdf, where each pdf element (text, paragraph, header, footer, chart etc.) detection is represented in a json format. Each element is denoted by its type, followed by its bounding box coordinates, followed by its value and/or its subelements. Bounding box coordinates are represented as " + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "inline_equation", + "content": "(x1, y1, x2, y2)" + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": ". These values correspond to the top left " + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "inline_equation", + "content": "x" + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": ", top left " + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "inline_equation", + "content": "y" + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": ", bottom right " + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "inline_equation", + "content": "x" + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": ", and bottom right " + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "inline_equation", + "content": "y" + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": ". The subelements are formatted and indented with spaces. The content of a pdf element is found at the 'value' key. The pdf elements are organized in the order in which they are present in the pdf: top to bottom, left to right. The pdfs can be single column, double column or multiple columns. Judge the number of columns in the pdfs by looking at the relative positioning of the pdf elements' " + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "inline_equation", + "content": "x1" + }, + { + "bbox": [ + 304, + 477, + 524, + 640 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 645, + 523, + 703 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 645, + 523, + 703 + ], + "spans": [ + { + "bbox": [ + 304, + 645, + 523, + 703 + ], + "type": "text", + "content": "Using the provided texts and coordinates, design question and answer pairs simulating the interactions between a user and system. Conversations should focus on potential user edit requests on the pdf elements (vs perception). Please follow the guidelines below while generating the conversations:" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 708, + 522, + 773 + ], + "type": "list", + "angle": 0, + "index": 22, + "blocks": [ + { + "bbox": [ + 304, + 708, + 522, + 742 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 708, + 522, + 742 + ], + "spans": [ + { + "bbox": [ + 304, + 708, + 522, + 742 + ], + "type": "text", + "content": "1. The edit requests should be diverse, requiring different kind of editing like adding, deleting, changing, modifying, swapping, moving, replacing, merging, splitting, and so on." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 749, + 502, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 749, + 502, + 773 + ], + "spans": [ + { + "bbox": [ + 304, + 749, + 502, + 773 + ], + "type": "text", + "content": "2. Please refer to the pdf-element that needs editing in diverse compositional ways. For example, " + }, + { + "bbox": [ + 304, + 749, + 502, + 773 + ], + "type": "inline_equation", + "content": "\\backslash" + }, + { + "bbox": [ + 304, + 749, + 502, + 773 + ], + "type": "text", + "content": " delete last line of the third" + } + ] + } + ], + "index": 21 + } + ], + "sub_type": "text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31133" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 6 + }, + { + "para_blocks": [ + { + "bbox": [ + 86, + 84, + 126, + 91 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 86, + 84, + 126, + 91 + ], + "spans": [ + { + "bbox": [ + 86, + 84, + 126, + 91 + ], + "type": "text", + "content": "System Prompt" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "spans": [ + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "content": "You are an AI visual assistance that can analyze PDFs. You will receive information describing a pdf, where each pdf element (text, paragraph, header, footer, chart etc.) detection is represented in a json format. Each element is denoted by its type, followed by its bounding box coordinates, followed by its value and/or its subelements. Bounding box coordinates are represented as (x1, y1, x2, y2). These values correspond to the top left " + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "inline_equation", + "content": "x" + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "content": ", top left " + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "inline_equation", + "content": "y" + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "content": ", bottom right " + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "inline_equation", + "content": "x" + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "content": ", and bottom right " + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "inline_equation", + "content": "y" + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "content": ". The subelements are formatted and indented with spaces. The content of a pdf element is found at the 'value' key. The pdf elements are organized in the order in which they are present in the pdf: top to bottom, left to right. The pdfs can be single column, double column or multiple columns. Judge the number of columns in the pdfs by looking at the relative positioning of the pdf elements' " + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "inline_equation", + "content": "x_1" + }, + { + "bbox": [ + 86, + 92, + 228, + 199 + ], + "type": "text", + "content": ". Using the provided texts and coordinates, design question and answer pairs simulating the interactions between a user and system. Conversations should focus on potential user edit requests on the pdf elements (vs perception)." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 86, + 203, + 210, + 215 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 86, + 203, + 210, + 215 + ], + "spans": [ + { + "bbox": [ + 86, + 203, + 210, + 215 + ], + "type": "text", + "content": "Please follow the guidelines below while generating the conversations:" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 86, + 216, + 227, + 258 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 86, + 216, + 227, + 258 + ], + "spans": [ + { + "bbox": [ + 86, + 216, + 227, + 258 + ], + "type": "text", + "content": "1. The edit requests should be diverse, requiring different kind of editing like adding, deleting, changing, modifying, swapping, moving, replacing, merging, splitting, and so on. \n2. Please refer to the pdf-element that needs editing in diverse compositional ways. For example, \"delete last line of the third paragraph\" or \"move the table in between the two figures from top to bottom\"." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 234, + 74, + 279, + 80 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 234, + 74, + 279, + 80 + ], + "spans": [ + { + "bbox": [ + 234, + 74, + 279, + 80 + ], + "type": "text", + "content": "One Shot Example" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 236, + 81, + 250, + 86 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 81, + 250, + 86 + ], + "spans": [ + { + "bbox": [ + 236, + 81, + 250, + 86 + ], + "type": "text", + "content": "" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 236, + 87, + 266, + 91 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 87, + 266, + 91 + ], + "spans": [ + { + "bbox": [ + 236, + 87, + 266, + 91 + ], + "type": "text", + "content": "element':text" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 236, + 92, + 287, + 97 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 92, + 287, + 97 + ], + "spans": [ + { + "bbox": [ + 236, + 92, + 287, + 97 + ], + "type": "text", + "content": "'bbox': [84, 382, 484, 413]," + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 236, + 98, + 308, + 103 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 98, + 308, + 103 + ], + "spans": [ + { + "bbox": [ + 236, + 98, + 308, + 103 + ], + "type": "text", + "content": "subelements: [element: line," + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 257, + 104, + 367, + 114 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 257, + 104, + 367, + 109 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 104, + 367, + 109 + ], + "spans": [ + { + "bbox": [ + 257, + 104, + 367, + 109 + ], + "type": "text", + "content": "'value' :interactions were then functionally annotated for" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 257, + 110, + 308, + 114 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 110, + 308, + 114 + ], + "spans": [ + { + "bbox": [ + 257, + 110, + 308, + 114 + ], + "type": "text", + "content": "value : Interactions were the" + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 255, + 116, + 286, + 121 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 255, + 116, + 286, + 121 + ], + "spans": [ + { + "bbox": [ + 255, + 116, + 286, + 121 + ], + "type": "text", + "content": "element: line" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 255, + 121, + 308, + 126 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 255, + 121, + 308, + 126 + ], + "spans": [ + { + "bbox": [ + 255, + 121, + 308, + 126 + ], + "type": "text", + "content": "\"Bbox:[84,397,377,414]" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 255, + 126, + 364, + 132 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 255, + 126, + 364, + 132 + ], + "spans": [ + { + "bbox": [ + 255, + 126, + 364, + 132 + ], + "type": "text", + "content": "'value': 'functional processes and pathways involved.']." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 236, + 132, + 287, + 137 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 132, + 287, + 137 + ], + "spans": [ + { + "bbox": [ + 236, + 132, + 287, + 137 + ], + "type": "text", + "content": "element':text'" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 236, + 137, + 287, + 142 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 137, + 287, + 142 + ], + "spans": [ + { + "bbox": [ + 236, + 137, + 287, + 142 + ], + "type": "text", + "content": "'bbox': [84, 410, 484, 498]," + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 255, + 143, + 297, + 148 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 255, + 143, + 297, + 148 + ], + "spans": [ + { + "bbox": [ + 255, + 143, + 297, + 148 + ], + "type": "text", + "content": "ents': ['element': 'line'," + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 255, + 148, + 313, + 153 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 255, + 148, + 313, + 153 + ], + "spans": [ + { + "bbox": [ + 255, + 148, + 313, + 153 + ], + "type": "text", + "content": "'bbox': [109, 411, 484, 428],..." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 236, + 153, + 255, + 158 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 153, + 255, + 158 + ], + "spans": [ + { + "bbox": [ + 236, + 153, + 255, + 158 + ], + "type": "text", + "content": "system>" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 234, + 158, + 369, + 179 + ], + "type": "list", + "angle": 0, + "index": 22, + "blocks": [ + { + "bbox": [ + 234, + 158, + 369, + 169 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 234, + 158, + 369, + 169 + ], + "spans": [ + { + "bbox": [ + 234, + 158, + 369, + 169 + ], + "type": "text", + "content": "User: Locate where changes are required to perform this query, 'Move the top figure from the center of the page to left'." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 234, + 169, + 369, + 179 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 234, + 169, + 369, + 179 + ], + "spans": [ + { + "bbox": [ + 234, + 169, + 369, + 179 + ], + "type": "text", + "content": "System: The figure that is required to move is located at [241, 88, 756, 323]." + } + ] + } + ], + "index": 21 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 234, + 184, + 368, + 204 + ], + "type": "list", + "angle": 0, + "index": 25, + "blocks": [ + { + "bbox": [ + 234, + 184, + 359, + 193 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 234, + 184, + 359, + 193 + ], + "spans": [ + { + "bbox": [ + 234, + 184, + 359, + 193 + ], + "type": "text", + "content": "User: Find the specific location requiring changes to perform this request. \"Underline the title.\" Results and Discussion" + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 234, + 194, + 368, + 204 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 234, + 194, + 368, + 204 + ], + "spans": [ + { + "bbox": [ + 234, + 194, + 368, + 204 + ], + "type": "text", + "content": "System: The title \"Results and Discussion\" can be found at [513, 675, 721, 691]." + } + ] + } + ], + "index": 24 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 234, + 207, + 261, + 213 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 234, + 207, + 261, + 213 + ], + "spans": [ + { + "bbox": [ + 234, + 207, + 261, + 213 + ], + "type": "text", + "content": "Detections" + } + ] + } + ], + "index": 26 + }, + { + "bbox": [ + 236, + 214, + 250, + 219 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 214, + 250, + 219 + ], + "spans": [ + { + "bbox": [ + 236, + 214, + 250, + 219 + ], + "type": "text", + "content": "'eleme" + } + ] + } + ], + "index": 27 + }, + { + "bbox": [ + 236, + 220, + 282, + 224 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 220, + 282, + 224 + ], + "spans": [ + { + "bbox": [ + 236, + 220, + 282, + 224 + ], + "type": "text", + "content": "'bbox': [76, 76, 179, 95]," + } + ] + } + ], + "index": 28 + }, + { + "bbox": [ + 236, + 224, + 250, + 228 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 224, + 250, + 228 + ], + "spans": [ + { + "bbox": [ + 236, + 224, + 250, + 228 + ], + "type": "text", + "content": "'subele" + } + ] + } + ], + "index": 29 + }, + { + "bbox": [ + 257, + 230, + 286, + 234 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 230, + 286, + 234 + ], + "spans": [ + { + "bbox": [ + 257, + 230, + 286, + 234 + ], + "type": "text", + "content": "'element': 'line'" + } + ] + } + ], + "index": 30 + }, + { + "bbox": [ + 257, + 235, + 305, + 240 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 235, + 305, + 240 + ], + "spans": [ + { + "bbox": [ + 257, + 235, + 305, + 240 + ], + "type": "text", + "content": "'bbox': [76, 76, 179, 94]," + } + ] + } + ], + "index": 31 + }, + { + "bbox": [ + 257, + 240, + 300, + 244 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 240, + 300, + 244 + ], + "spans": [ + { + "bbox": [ + 257, + 240, + 300, + 244 + ], + "type": "text", + "content": "value':AppendDIX}\\}}," + } + ] + } + ], + "index": 32 + }, + { + "bbox": [ + 236, + 245, + 246, + 249 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 245, + 246, + 249 + ], + "spans": [ + { + "bbox": [ + 236, + 245, + 246, + 249 + ], + "type": "text", + "content": "'elem" + } + ] + } + ], + "index": 33 + }, + { + "bbox": [ + 236, + 250, + 287, + 255 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 250, + 287, + 255 + ], + "spans": [ + { + "bbox": [ + 236, + 250, + 287, + 255 + ], + "type": "text", + "content": "'bbox': [76, 123, 489, 183]" + } + ] + } + ], + "index": 34 + }, + { + "bbox": [ + 236, + 255, + 267, + 259 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 255, + 267, + 259 + ], + "spans": [ + { + "bbox": [ + 236, + 255, + 267, + 259 + ], + "type": "text", + "content": "'subelements': I" + } + ] + } + ], + "index": 35 + }, + { + "bbox": [ + 257, + 260, + 332, + 275 + ], + "type": "list", + "angle": 0, + "index": 39, + "blocks": [ + { + "bbox": [ + 257, + 260, + 287, + 264 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 260, + 287, + 264 + ], + "spans": [ + { + "bbox": [ + 257, + 260, + 287, + 264 + ], + "type": "text", + "content": "element: line." + } + ] + } + ], + "index": 36 + }, + { + "bbox": [ + 257, + 265, + 308, + 269 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 265, + 308, + 269 + ], + "spans": [ + { + "bbox": [ + 257, + 265, + 308, + 269 + ], + "type": "text", + "content": "'bbox':[76,122,485,139]" + } + ] + } + ], + "index": 37 + }, + { + "bbox": [ + 257, + 270, + 332, + 275 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 257, + 270, + 332, + 275 + ], + "spans": [ + { + "bbox": [ + 257, + 270, + 332, + 275 + ], + "type": "text", + "content": "'value': 'As stated in Section 2.1, the..." + } + ] + } + ], + "index": 38 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 404, + 74, + 508, + 277 + ], + "type": "list", + "angle": 0, + "index": 52, + "blocks": [ + { + "bbox": [ + 404, + 74, + 508, + 91 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 404, + 74, + 508, + 91 + ], + "spans": [ + { + "bbox": [ + 404, + 74, + 508, + 91 + ], + "type": "text", + "content": "User: Locate where changes are required to perform this query, 'Add \"Introduction\" as a new title before \"Appendix\"." + } + ] + } + ], + "index": 40 + }, + { + "bbox": [ + 405, + 95, + 508, + 105 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 95, + 508, + 105 + ], + "spans": [ + { + "bbox": [ + 405, + 95, + 508, + 105 + ], + "type": "text", + "content": "System: The location where \"Appendix\" is found is at [76, 76, 179, 95]." + } + ] + } + ], + "index": 41 + }, + { + "bbox": [ + 405, + 116, + 506, + 132 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 116, + 506, + 132 + ], + "spans": [ + { + "bbox": [ + 405, + 116, + 506, + 132 + ], + "type": "text", + "content": "User: Find the specific location requiring changes to perform this request, 'Delete the third list item under \"A.1.Environment\":" + } + ] + } + ], + "index": 42 + }, + { + "bbox": [ + 405, + 136, + 506, + 148 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 136, + 506, + 148 + ], + "spans": [ + { + "bbox": [ + 405, + 136, + 506, + 148 + ], + "type": "text", + "content": "System: The third list item under \"A.1.Environment\" can be found at [92, 398, 485, 459]." + } + ] + } + ], + "index": 43 + }, + { + "bbox": [ + 405, + 157, + 498, + 162 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 157, + 498, + 162 + ], + "spans": [ + { + "bbox": [ + 405, + 157, + 498, + 162 + ], + "type": "text", + "content": "User: Find the place that needs to be edited to" + } + ] + } + ], + "index": 44 + }, + { + "bbox": [ + 405, + 163, + 499, + 168 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 163, + 499, + 168 + ], + "spans": [ + { + "bbox": [ + 405, + 163, + 499, + 168 + ], + "type": "text", + "content": "process this query, 'Move the list under \"A.1.4.\"" + } + ] + } + ], + "index": 45 + }, + { + "bbox": [ + 405, + 169, + 498, + 174 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 169, + 498, + 174 + ], + "spans": [ + { + "bbox": [ + 405, + 169, + 498, + 174 + ], + "type": "text", + "content": "\"Brain-Body Interface\" to the bottom of the page!" + } + ] + } + ], + "index": 46 + }, + { + "bbox": [ + 405, + 178, + 506, + 189 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 178, + 506, + 189 + ], + "spans": [ + { + "bbox": [ + 405, + 178, + 506, + 189 + ], + "type": "text", + "content": "System: The list under \"A.1.4. Brain-Body Interface\" can be found at [514, 175, 926, 531]." + } + ] + } + ], + "index": 47 + }, + { + "bbox": [ + 405, + 194, + 507, + 209 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 194, + 507, + 209 + ], + "spans": [ + { + "bbox": [ + 405, + 194, + 507, + 209 + ], + "type": "text", + "content": "User: Please find the area that needs editing to carry out this user request, 'Merge the two list items under \"A.1.5. Experiment\"." + } + ] + } + ], + "index": 48 + }, + { + "bbox": [ + 405, + 214, + 507, + 230 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 214, + 507, + 230 + ], + "spans": [ + { + "bbox": [ + 405, + 214, + 507, + 230 + ], + "type": "text", + "content": "System: The two list items under \"A.1.5. Experiment\" can be found at [530, 568, 923, 658] and [530, 657, 923, 806]." + } + ] + } + ], + "index": 49 + }, + { + "bbox": [ + 405, + 240, + 507, + 256 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 240, + 507, + 256 + ], + "spans": [ + { + "bbox": [ + 405, + 240, + 507, + 256 + ], + "type": "text", + "content": "User: Please locate the part that needs modification to process this query, \"Swap the order of the first two list items under \"A.1. Design and Editing\"." + } + ] + } + ], + "index": 50 + }, + { + "bbox": [ + 405, + 260, + 506, + 277 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 405, + 260, + 506, + 277 + ], + "spans": [ + { + "bbox": [ + 405, + 260, + 506, + 277 + ], + "type": "text", + "content": "System: The first two list items under \"A.1. Design and Editing\" can be found at [92, 236, 485, 341] and [92, 339, 485, 400]." + } + ] + } + ], + "index": 51 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 122, + 290, + 468, + 303 + ], + "angle": 0, + "lines": [ + { + "bbox": [ + 122, + 290, + 468, + 303 + ], + "spans": [ + { + "bbox": [ + 122, + 290, + 468, + 303 + ], + "type": "text", + "content": "Figure 7: Illustration of LLM prompting to generate synthetic instruction tuning data." + } + ] + } + ], + "index": 53, + "type": "text" + }, + { + "bbox": [ + 89, + 319, + 287, + 336 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 319, + 287, + 336 + ], + "spans": [ + { + "bbox": [ + 89, + 319, + 287, + 336 + ], + "type": "text", + "content": "paragraph\\\" or \\\\\" move the table in between the two figures from top to bottom \\\"." + } + ] + } + ], + "index": 54 + }, + { + "bbox": [ + 69, + 343, + 287, + 758 + ], + "type": "list", + "angle": 0, + "index": 67, + "blocks": [ + { + "bbox": [ + 69, + 343, + 271, + 367 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 343, + 271, + 367 + ], + "spans": [ + { + "bbox": [ + 69, + 343, + 271, + 367 + ], + "type": "text", + "content": "3. Please emphasize edit requests that deal with more abstract pdf elements rather low level elements like words and lines." + } + ] + } + ], + "index": 55 + }, + { + "bbox": [ + 69, + 375, + 279, + 398 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 375, + 279, + 398 + ], + "spans": [ + { + "bbox": [ + 69, + 375, + 279, + 398 + ], + "type": "text", + "content": "4. Use the x and y coordinates of the pdf-elements to figure out the relative spatial relations left of, right of." + } + ] + } + ], + "index": 56 + }, + { + "bbox": [ + 69, + 400, + 279, + 414 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 400, + 279, + 414 + ], + "spans": [ + { + "bbox": [ + 69, + 400, + 279, + 414 + ], + "type": "text", + "content": "top, bottom or to assess which paragraph is first, second or third." + } + ] + } + ], + "index": 57 + }, + { + "bbox": [ + 69, + 423, + 287, + 454 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 423, + 287, + 454 + ], + "spans": [ + { + "bbox": [ + 69, + 423, + 287, + 454 + ], + "type": "text", + "content": "5. Please count the number and order of pdf elements sequentially and correctly. Mistakes in counting are costly. Please think step by step while doing it." + } + ] + } + ], + "index": 58 + }, + { + "bbox": [ + 69, + 462, + 287, + 502 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 462, + 287, + 502 + ], + "spans": [ + { + "bbox": [ + 69, + 462, + 287, + 502 + ], + "type": "text", + "content": "6. Please do not generate edit requests that implies modifying multiple pdf elements when there exists actually only one, for example, asking to swap the order of list-items when acutely there is a single list item." + } + ] + } + ], + "index": 59 + }, + { + "bbox": [ + 69, + 510, + 284, + 534 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 510, + 284, + 534 + ], + "spans": [ + { + "bbox": [ + 69, + 510, + 284, + 534 + ], + "type": "text", + "content": "7. Please do not generate edits that call for left/right column, where there exists only a single column." + } + ] + } + ], + "index": 60 + }, + { + "bbox": [ + 69, + 542, + 287, + 591 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 542, + 287, + 591 + ], + "spans": [ + { + "bbox": [ + 69, + 542, + 287, + 591 + ], + "type": "text", + "content": "8. In the presence of list, please generate at least one edit requests on list bullets. The output then should also refer only to the bullets, not the whole list. Check to see whether the bullets are numeric, alphabetic, or plain circular bullets before requesting its editing." + } + ] + } + ], + "index": 61 + }, + { + "bbox": [ + 69, + 598, + 286, + 622 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 598, + 286, + 622 + ], + "spans": [ + { + "bbox": [ + 69, + 598, + 286, + 622 + ], + "type": "text", + "content": "9. Please geenrate the minimal bounding box that suffices to make the edit, i.e. ground the user request to the most fine-grained pdf-element." + } + ] + } + ], + "index": 62 + }, + { + "bbox": [ + 69, + 629, + 286, + 677 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 629, + 286, + 677 + ], + "spans": [ + { + "bbox": [ + 69, + 629, + 286, + 677 + ], + "type": "text", + "content": "10. Please generate multiple back and forth conversations. The user query should follow this format: \", \\\"\\\". For example: 'Please find the location to make this edit, \\ \"Change list bullets to numeric\\\"." + } + ] + } + ], + "index": 63 + }, + { + "bbox": [ + 69, + 685, + 286, + 710 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 685, + 286, + 710 + ], + "spans": [ + { + "bbox": [ + 69, + 685, + 286, + 710 + ], + "type": "text", + "content": "11. Only generate questions and answers that can be answered with the given information, do not use background information." + } + ] + } + ], + "index": 64 + }, + { + "bbox": [ + 69, + 717, + 267, + 733 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 717, + 267, + 733 + ], + "spans": [ + { + "bbox": [ + 69, + 717, + 267, + 733 + ], + "type": "text", + "content": "12. Please do not hallucinate pdf elements, its content, its subelements or edit requests." + } + ] + } + ], + "index": 65 + }, + { + "bbox": [ + 69, + 740, + 284, + 758 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 740, + 284, + 758 + ], + "spans": [ + { + "bbox": [ + 69, + 740, + 284, + 758 + ], + "type": "text", + "content": "13. It's better to not generate requests instead of generating wrong requests." + } + ] + } + ], + "index": 66 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 303, + 315, + 489, + 330 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 315, + 489, + 330 + ], + "spans": [ + { + "bbox": [ + 303, + 315, + 489, + 330 + ], + "type": "text", + "content": "C Code-NLI Prompt and Samples" + } + ] + } + ], + "index": 68 + }, + { + "bbox": [ + 302, + 337, + 526, + 405 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 337, + 526, + 405 + ], + "spans": [ + { + "bbox": [ + 302, + 337, + 526, + 405 + ], + "type": "text", + "content": "To clean noisy synthetic data, we employ Code-NLI. The prompt used for Code-NLI includes a system prompt, that contains Python API function definitions, and few-shot prompts. We provide the system prompt given below:" + } + ] + } + ], + "index": 69 + }, + { + "bbox": [ + 304, + 414, + 523, + 774 + ], + "type": "list", + "angle": 0, + "index": 74, + "blocks": [ + { + "bbox": [ + 304, + 414, + 523, + 503 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 414, + 523, + 503 + ], + "spans": [ + { + "bbox": [ + 304, + 414, + 523, + 503 + ], + "type": "text", + "content": "You are an AI visual assistant that can do Natural Language Inference using python code. You will receive information describing a hypothesis. This includes a user edit request on pdf and a system output specifying the edit location with a bounding box. Your job is to create a python code, using listed APIs, that verifies that the bounding box correctly localizes the edit target in the user request. Each statement in the python code needs to be true to verify the correctness of the hypothesis." + } + ] + } + ], + "index": 70 + }, + { + "bbox": [ + 304, + 510, + 523, + 631 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 510, + 523, + 631 + ], + "spans": [ + { + "bbox": [ + 304, + 510, + 523, + 631 + ], + "type": "text", + "content": "The user edit request focuses on editing different type of pdf elements: title, paragraph, list, figure, and table. The elements could have subelements: title -> {bullets -> {bullet}, list_item -> {line}}, paragraph -> {line}, title -> {line}. It refers to different elements in a compositional and hierarchical manner. For example, last line of second paragraph, first list-item of third list, etc. The system output localizes the element/subelement requiring edit with a bounding box. The format of the bounding box is [x1,y1,x2,y2] where (x1,y1) is the coordinate of top left corner and (x2,y2) is bottom right corner, where x2>x1 and y2>y1." + } + ] + } + ], + "index": 71 + }, + { + "bbox": [ + 304, + 637, + 523, + 702 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 637, + 523, + 702 + ], + "spans": [ + { + "bbox": [ + 304, + 637, + 523, + 702 + ], + "type": "text", + "content": "Please use the following APIs to verify that the output bounding box indeed correctly captures the edit element/subelement. Please output only the python code using the listed APIs and nothing else. The element (dict) follows this format: {element_type: ,_bbox: []}, value: , subelements:[subelement1,..]}." + } + ] + } + ], + "index": 72 + }, + { + "bbox": [ + 304, + 708, + 507, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 708, + 507, + 774 + ], + "spans": [ + { + "bbox": [ + 304, + 708, + 507, + 774 + ], + "type": "text", + "content": "def is_element.element_type: str, bbox: list) -> bool:\n \"checks whether the element in_bbox is element_type\"\n \"Args: element_type (str): any element type ' title', 'paragraph', 'list', 'line', 'figure', 'table', 'bullets', '" + } + ] + } + ], + "index": 73 + } + ], + "sub_type": "text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31134" + } + ] + } + ], + "index": 75 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 7 + }, + { + "para_blocks": [ + { + "type": "code", + "bbox": [ + 65, + 73, + 296, + 769 + ], + "blocks": [ + { + "bbox": [ + 65, + 73, + 296, + 769 + ], + "lines": [ + { + "bbox": [ + 65, + 73, + 296, + 769 + ], + "spans": [ + { + "bbox": [ + 65, + 73, + 296, + 769 + ], + "type": "text", + "content": "def in_sectionsection section_title: str, bbox: list) -> bool:\n '''checks whether the bbox is in the section_title''' \n '''Args: section_title (str): the string contained in the title \n bbox (list): a bounding box of the type [ x1, y1, x2, y2] \n ''' \ndef contains_string (string: str,_bbox: list) -> bool:\n '''checks whether the element in_bbox contains string''' \n '''Args: string (str): string to check \n _bbox (list): a bounding box of the type [ x1, y1, x2, y2] \n ''' \ndef check_element_order_from_top_or_title( top_or_title (str), element_type: str,_bbox: \n list, order: int, column: str = None) -> bool: \n '''checks the sequence order of the element in \n _bbox of element_type from top of pdf or the section title where it lies''' \n '''Args: \n top_or_title (str): Either 'top' or 'title' from where to beign counting \n element_type (str): First hierarchy elements -- 'title', 'paragraph', 'list', 'figure' or 'table' \n _bbox (list): a bounding box of the type [ x1, y1, x2, y2] \n order (int): denoting the sequence order of bounding box of element_type. 1,2,3 order counts from top, -1,-2.. order counts from last \n column (str): to check the order in left or right column. Defaults to None when pdf is single column. \n */\ndef check_subelement_order_from_top_or_title( top_or_title (str), subelement_type: str,_bbox: \n list, order: int, column: str = None) -> int: \n '''checks the sequence order of the subelement in bounding box of subelement_type from top of pdf or the section title where it lies''' \n '''Args: \n top_or_title (str): Either 'top' or 'title' from where to beign counting \n subelement_type (str): Second hierarchy elements -- 'line' (title and paragraph subelement), \\ 'bullets' (list subelement), '\\list_item' (list subelement) \n _bbox (list): a bounding box of the type [ x1, y1, x2, y2] \n order (int): denoting the sequence order of bounding box of subelement_type. 1,2,3 order counts from top, -1,-2.. order counts from last \n column (str): to check the order in left or right column. Defaults to None when pdf is single column. \n */\ndef check_subelement_order_in_element( subelement_type: str,_bbox: list, order: int) -> int: \n '''checks the sequence order of the subelement in bounding box of subelement_type within the element in which_bbox lies''' \n '''Args: \n subelement_type (str): Second hierarchy elements -- 'line' (title and paragraph subelement), 'bullets' (list subelement), 'list_item' (list subelement) \n _bbox (list): a bounding box of the type [" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "code_body" + } + ], + "index": 0, + "sub_type": "code", + "guess_lang": "python" + }, + { + "type": "code", + "bbox": [ + 298, + 73, + 530, + 497 + ], + "blocks": [ + { + "bbox": [ + 298, + 73, + 530, + 497 + ], + "lines": [ + { + "bbox": [ + 298, + 73, + 530, + 497 + ], + "spans": [ + { + "bbox": [ + 298, + 73, + 530, + 497 + ], + "type": "text", + "content": "x1, y1, x2, y2]\norder (int): denoting the sequence order of_bbox of subelement_type. 1,2,3\norder counts from top, -1,-2.. order counts from last\n\" \"\ndef check_subsubelement_order_from_top_or_title(top_or_title (str), subsublement_type: str,_bbox: list, order: int, column: str = None) -> int:\n \" checks the sequence order of the subsublement in_bbox of subsublement_type from top of pdf or the section title where it lies\"\n \"Args:\n top_or_title (str): Either 'top' or 'title' from where to beign counting subsublement_type (str): Third hierarchy elements -- 'bullet' (subsublement of bullets), 'line' (subsublement of list_item)\nbbox (list): a bounding box of the type [x1, y1, x2, y2]\norder (int): denoting the sequence order of_bbox of subelement_type. 1,2,3 order counts from top, -1,-2.. order counts from last column (str): to check the order in left or right column. Defaults to None when pdf is single column.\n\" \"\ndef check_subsubelement_order_in_subelement(subsublement_type: str,_bbox: list, order: int) -> int:\n \" checks the sequence order of the subsublement in_bbox of subsublement_type within the subelement in which_bbox lies\"\n \"Args:\n subsublement_type (str): Third hierarchy elements -- 'bullet' (subsublement of bullets), 'line' (subsublement of list_item)\nbbox (list): a bounding box of the type [x1, y1, x2, y2]\norder (int): denoting the sequence order of_bbox of subsublement_type. 1,2,3 order counts from top, -1,-2.. order counts from last" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "code_body" + } + ], + "index": 1, + "sub_type": "code", + "guess_lang": "python" + }, + { + "bbox": [ + 302, + 502, + 526, + 529 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 502, + 526, + 529 + ], + "spans": [ + { + "bbox": [ + 302, + 502, + 526, + 529 + ], + "type": "text", + "content": "We also include some sample Python code generated by Code-NLI for verification in Figure 8." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 303, + 540, + 415, + 553 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 540, + 415, + 553 + ], + "spans": [ + { + "bbox": [ + 303, + 540, + 415, + 553 + ], + "type": "text", + "content": "D Dataset Statistics" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "spans": [ + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "type": "text", + "content": "The total size of the synthetically generated data is 3M samples. The topic range is the same as the topic distribution of the base datasets: scientific articles, patents, finance, tenders, laws, and manuals. Similarly, image resolution is the same as base datasets: " + }, + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "type": "inline_equation", + "content": "1025 \\times 1025" + }, + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "type": "text", + "content": " for DocLayNet and " + }, + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "type": "inline_equation", + "content": "800 \\times 600" + }, + { + "bbox": [ + 302, + 564, + 526, + 658 + ], + "type": "text", + "content": " for PubLayNet." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 303, + 671, + 427, + 683 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 671, + 427, + 683 + ], + "spans": [ + { + "bbox": [ + 303, + 671, + 427, + 683 + ], + "type": "text", + "content": "E Model Architecture" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 693, + 526, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 693, + 526, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 693, + 526, + 775 + ], + "type": "text", + "content": ") Our model, DELOC, is based on Phi-3-V model. The Phi-3-V model's architecture is essentially composed of two components - an image encoder (CLIP ViT-L/14 (Radford et al., 2021)) and a transformer decoder (phi-3-mini). The visual tokens from the image encoder and concatenate with text" + } + ] + } + ], + "index": 6 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31135" + } + ] + } + ], + "index": 7 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 8 + }, + { + "para_blocks": [ + { + "bbox": [ + 87, + 74, + 263, + 107 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 87, + 74, + 263, + 107 + ], + "spans": [ + { + "bbox": [ + 87, + 74, + 263, + 107 + ], + "type": "text", + "content": "User: Please locate the part that needs to be edited to process this query, 'Delete the last line of the last paragraph in the left column'." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 87, + 116, + 265, + 148 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 87, + 116, + 265, + 148 + ], + "spans": [ + { + "bbox": [ + 87, + 116, + 265, + 148 + ], + "type": "text", + "content": "System: The last paragraph in the left column is located at [84, 654, 484, 813] and the last line is at [84, 797, 140, 814]." + } + ] + } + ], + "index": 1 + }, + { + "type": "image", + "bbox": [ + 274, + 95, + 315, + 132 + ], + "blocks": [ + { + "bbox": [ + 274, + 95, + 315, + 132 + ], + "lines": [ + { + "bbox": [ + 274, + 95, + 315, + 132 + ], + "spans": [ + { + "bbox": [ + 274, + 95, + 315, + 132 + ], + "type": "image", + "image_path": "a7ca97b45a5e4edf62988dd8e494e6dcc3ce4a209cc0416d0e09faa1672a2d20.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_body" + } + ], + "index": 2 + }, + { + "bbox": [ + 318, + 74, + 484, + 149 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 318, + 74, + 484, + 149 + ], + "spans": [ + { + "bbox": [ + 318, + 74, + 484, + 149 + ], + "type": "text", + "content": "is_element('list_item', [92,398,485,459]) \nin_section('A.1.Environment', [92,398,485,459]) \ncheck_subelement_order_in_element('list_item', [92,398,485,459], 2)" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 87, + 169, + 265, + 211 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 87, + 169, + 265, + 211 + ], + "spans": [ + { + "bbox": [ + 87, + 169, + 265, + 211 + ], + "type": "text", + "content": "User: Please find the location that needs to be edited to process this query, 'Delete the reference \"[23]\" from the second paragraph in the left column'." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 87, + 222, + 258, + 254 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 87, + 222, + 258, + 254 + ], + "spans": [ + { + "bbox": [ + 87, + 222, + 258, + 254 + ], + "type": "text", + "content": "System: The second paragraph in the left column is located at [84, 410, 484, 498] and the reference \"[23]\" is at [84, 482, 167, 499]." + } + ] + } + ], + "index": 5 + }, + { + "type": "image", + "bbox": [ + 272, + 191, + 315, + 227 + ], + "blocks": [ + { + "bbox": [ + 272, + 191, + 315, + 227 + ], + "lines": [ + { + "bbox": [ + 272, + 191, + 315, + 227 + ], + "spans": [ + { + "bbox": [ + 272, + 191, + 315, + 227 + ], + "type": "image", + "image_path": "6937a613b7a1c3104cba624cf2eed197664f770b1ee7fcbcd80f1db17754f3be.jpg" + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "image_body" + } + ], + "index": 6 + }, + { + "bbox": [ + 319, + 169, + 478, + 254 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 319, + 169, + 478, + 254 + ], + "spans": [ + { + "bbox": [ + 319, + 169, + 478, + 254 + ], + "type": "text", + "content": "is_element('paragraph', [84, 410, 484, 498]) \ncheck_element_order_from_top_or_title('top', 'paragraph', [84, 410, 484, 498], 2, column='left') \ncontains_string(['[23]', [84, 482, 167, 499])" + } + ] + } + ], + "index": 7 + }, + { + "type": "image", + "bbox": [ + 75, + 289, + 205, + 468 + ], + "blocks": [ + { + "bbox": [ + 174, + 264, + 417, + 277 + ], + "lines": [ + { + "bbox": [ + 174, + 264, + 417, + 277 + ], + "spans": [ + { + "bbox": [ + 174, + 264, + 417, + 277 + ], + "type": "text", + "content": "Figure 8: Sample verification code generated by Code-NLI." + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 75, + 289, + 205, + 468 + ], + "lines": [ + { + "bbox": [ + 75, + 289, + 205, + 468 + ], + "spans": [ + { + "bbox": [ + 75, + 289, + 205, + 468 + ], + "type": "image", + "image_path": "aeac7283b8e342b7489f868ad18d6a6956cb6f467918627ef010a991abdef461.jpg" + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 75, + 473, + 204, + 505 + ], + "lines": [ + { + "bbox": [ + 75, + 473, + 204, + 505 + ], + "spans": [ + { + "bbox": [ + 75, + 473, + 204, + 505 + ], + "type": "text", + "content": "The length of the table \"Exhibit 2\" is increased. The spread of the table is increased towards the left margin.." + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_footnote" + } + ], + "index": 9 + }, + { + "type": "image", + "bbox": [ + 211, + 290, + 356, + 478 + ], + "blocks": [ + { + "bbox": [ + 211, + 290, + 356, + 478 + ], + "lines": [ + { + "bbox": [ + 211, + 290, + 356, + 478 + ], + "spans": [ + { + "bbox": [ + 211, + 290, + 356, + 478 + ], + "type": "image", + "image_path": "a97ad259172bccbbc8ad15338ef927fa6b6241ab34fece8369cc737a9f929eed.jpg" + } + ] + } + ], + "index": 11, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 215, + 481, + 351, + 503 + ], + "lines": [ + { + "bbox": [ + 215, + 481, + 351, + 503 + ], + "spans": [ + { + "bbox": [ + 215, + 481, + 351, + 503 + ], + "type": "text", + "content": "Moved the text \"Golden Chopstick, Takeaway menu\" above the purple line" + } + ] + } + ], + "index": 12, + "angle": 0, + "type": "image_footnote" + }, + { + "bbox": [ + 102, + 514, + 489, + 527 + ], + "lines": [ + { + "bbox": [ + 102, + 514, + 489, + 527 + ], + "spans": [ + { + "bbox": [ + 102, + 514, + 489, + 527 + ], + "type": "text", + "content": "Figure 9: Additional Qualitative sample prediction from DELOC demonstrating its capabilities." + } + ] + } + ], + "index": 15, + "angle": 0, + "type": "image_caption" + } + ], + "index": 11 + }, + { + "type": "image", + "bbox": [ + 360, + 290, + 520, + 473 + ], + "blocks": [ + { + "bbox": [ + 360, + 290, + 520, + 473 + ], + "lines": [ + { + "bbox": [ + 360, + 290, + 520, + 473 + ], + "spans": [ + { + "bbox": [ + 360, + 290, + 520, + 473 + ], + "type": "image", + "image_path": "00d5d6cd15d5ff3ebc3611e090cd1b138088bd89b6b1835a0a05884bb0263566.jpg" + } + ] + } + ], + "index": 13, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 411, + 492, + 521, + 504 + ], + "lines": [ + { + "bbox": [ + 411, + 492, + 521, + 504 + ], + "spans": [ + { + "bbox": [ + 411, + 492, + 521, + 504 + ], + "type": "text", + "content": "GT: Pred:" + } + ] + } + ], + "index": 14, + "angle": 0, + "type": "image_footnote" + } + ], + "index": 13 + }, + { + "bbox": [ + 67, + 540, + 290, + 579 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 540, + 290, + 579 + ], + "spans": [ + { + "bbox": [ + 67, + 540, + 290, + 579 + ], + "type": "text", + "content": "tokens in an interleaved manner and fed to the transformer decoder to output the prediction. In total, the number of parameters totals 4.2B." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 67, + 581, + 289, + 608 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 581, + 289, + 608 + ], + "spans": [ + { + "bbox": [ + 67, + 581, + 289, + 608 + ], + "type": "text", + "content": "The Phi-3-mini is a transformer decoder with 3072 hidden dimension, 32 heads and 32 layers." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 67, + 618, + 230, + 631 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 618, + 230, + 631 + ], + "spans": [ + { + "bbox": [ + 67, + 618, + 230, + 631 + ], + "type": "text", + "content": "F Additional Training Details" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 67, + 640, + 291, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 640, + 291, + 774 + ], + "spans": [ + { + "bbox": [ + 67, + 640, + 291, + 774 + ], + "type": "text", + "content": "The input image to our model is resized to " + }, + { + "bbox": [ + 67, + 640, + 291, + 774 + ], + "type": "inline_equation", + "content": "900 \\times 900" + }, + { + "bbox": [ + 67, + 640, + 291, + 774 + ], + "type": "text", + "content": ". The text-heavy nature of PDF images requires that the model handle image resolutions dynamically according to the resolution of the text. Higher resolution text should result in higher image input resolution, and vice versa. This is taken care of by the dynamic cropping strategy (Dong et al., 2024) in the Phi-3-V model. It allows the input image to be dynamically cropped into a 2d array of blocks, which are then concatenated to represent the whole" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 303, + 540, + 336, + 553 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 540, + 336, + 553 + ], + "spans": [ + { + "bbox": [ + 303, + 540, + 336, + 553 + ], + "type": "text", + "content": "image." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 302, + 564, + 495, + 577 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 564, + 495, + 577 + ], + "spans": [ + { + "bbox": [ + 302, + 564, + 495, + 577 + ], + "type": "text", + "content": "G Additional Qualitative Examples" + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 302, + 585, + 526, + 639 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 585, + 526, + 639 + ], + "spans": [ + { + "bbox": [ + 302, + 585, + 526, + 639 + ], + "type": "text", + "content": "We add more qualitative samples to further illustrate the capabilities of DELOC. As can be seen in Figure 9, DELOC is able to ground user edit requests well both for PDFs and design documents." + } + ] + } + ], + "index": 22 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "31136" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 9 + } + ], + "_backend": "vlm", + "_version_name": "2.6.4" +} \ No newline at end of file