diff --git "a/2023/A Survey of Deep Learning for Mathematical Reasoning/layout.json" "b/2023/A Survey of Deep Learning for Mathematical Reasoning/layout.json" new file mode 100644--- /dev/null +++ "b/2023/A Survey of Deep Learning for Mathematical Reasoning/layout.json" @@ -0,0 +1,16153 @@ +{ + "pdf_info": [ + { + "para_blocks": [ + { + "bbox": [ + 123, + 76, + 470, + 95 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 123, + 76, + 470, + 95 + ], + "spans": [ + { + "bbox": [ + 123, + 76, + 470, + 95 + ], + "type": "text", + "content": "A Survey of Deep Learning for Mathematical Reasoning" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "spans": [ + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": "Pan Lu" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": ", Liang Qiu" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": ", Wenhao Yu" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{2}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": ", Sean Welleck" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{3*}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": ", Kai-Wei Chang" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{1*}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": "UCLA, " + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{2}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": "University of Notre Dame, " + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "inline_equation", + "content": "^{3}" + }, + { + "bbox": [ + 121, + 118, + 474, + 163 + ], + "type": "text", + "content": "University of Washington https://github.com/lupantech/dl4math" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 155, + 212, + 202, + 226 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 155, + 212, + 202, + 226 + ], + "spans": [ + { + "bbox": [ + 155, + 212, + 202, + 226 + ], + "type": "text", + "content": "Abstract" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 86, + 233, + 274, + 496 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 86, + 233, + 274, + 496 + ], + "spans": [ + { + "bbox": [ + 86, + 233, + 274, + 496 + ], + "type": "text", + "content": "Mathematical reasoning is a fundamental aspect of human intelligence and is applicable in various fields, including science, engineering, finance, and everyday life. The development of artificial intelligence (AI) systems capable of solving math problems and proving theorems in language has garnered significant interest in the fields of machine learning and natural language processing. For example, mathematics serves as a testbed for aspects of reasoning that are challenging for powerful deep learning models, driving new algorithmic and modeling advances. On the other hand, recent advances in large-scale neural language models have opened up new benchmarks and opportunities to use deep learning for mathematical reasoning. In this survey paper, we review the key tasks, datasets, and methods at the intersection of mathematical reasoning and deep learning over the past decade. We also evaluate existing benchmarks and methods, and discuss future research directions in this domain." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 68, + 505, + 154, + 518 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 505, + 154, + 518 + ], + "spans": [ + { + "bbox": [ + 68, + 505, + 154, + 518 + ], + "type": "text", + "content": "1 Introduction" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 526, + 289, + 552 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 526, + 289, + 552 + ], + "spans": [ + { + "bbox": [ + 67, + 526, + 289, + 552 + ], + "type": "text", + "content": "\"The study of mathematics, like the Nile, begins in minuteness but ends in magnificence.\"" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 108, + 553, + 286, + 566 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 108, + 553, + 286, + 566 + ], + "spans": [ + { + "bbox": [ + 108, + 553, + 286, + 566 + ], + "type": "text", + "content": "— Charles Caleb Colton, English writer" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 580, + 291, + 756 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 580, + 291, + 756 + ], + "spans": [ + { + "bbox": [ + 67, + 580, + 291, + 756 + ], + "type": "text", + "content": "Mathematical reasoning is a key aspect of human intelligence that enables us to comprehend and make decisions based on numerical data and language. It is applicable in various fields, including science, engineering, finance, and everyday life, and encompasses a range of abilities, from basic skills such as pattern recognition and numerical operations to more advanced skills like problem-solving, logical reasoning, and abstract thinking. The development of artificial intelligence (AI) systems capable of solving math problems and proving theorems in language has been a long-standing focus of research in the fields of machine learning and" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 213, + 526, + 294 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 213, + 526, + 294 + ], + "spans": [ + { + "bbox": [ + 302, + 213, + 526, + 294 + ], + "type": "text", + "content": "natural language processing (NLP), dating back to the 1960s (Feigenbaum et al., 1963; Bobrow, 1964). In recent years, there has been a surge of interest in this area: for instance, the number of papers has grown from approximately 10 in 2018 to 66 in 2022 (see Figure 3 in the Appendix)." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 296, + 526, + 444 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 296, + 526, + 444 + ], + "spans": [ + { + "bbox": [ + 302, + 296, + 526, + 444 + ], + "type": "text", + "content": "As deep learning continues to revolutionize NLP tasks such as question answering and machine translation (Sutskever et al., 2014; Devlin et al., 2019), it has also made significant strides in the field of mathematical reasoning (Wang et al., 2017; Yang and Deng, 2019; Geva et al., 2020; Wei et al., 2022). However, despite the impressive capabilities of these models, there is still a lack of a clear taxonomy of the different types of mathematical reasoning tasks and the specific capabilities required of deep learning models to solve them." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 446, + 526, + 649 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 446, + 526, + 649 + ], + "spans": [ + { + "bbox": [ + 302, + 446, + 526, + 649 + ], + "type": "text", + "content": "Previous literature has been limited to the discussion of specific aspects, such as solving math word problems (Bhattacharya, 2017; Zhang et al., 2019; Ughade and Kumbhar, 2019), representing numbers representation (Thawani et al., 2021), or solving informal problems (Meadows and Freitas, 2022). Additionally, with the recent advancements in large language models like GPT-3 (Brown et al., 2020), there is a growing need to understand the capabilities and limitations of these models in the context of mathematical reasoning. This is where a comprehensive survey of this rapidly advancing domain becomes crucial, as it can provide an overview of the current state and limitations of the field, and indicate further research areas." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 651, + 526, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 651, + 526, + 773 + ], + "spans": [ + { + "bbox": [ + 302, + 651, + 526, + 773 + ], + "type": "text", + "content": "In this paper, we survey over 180 papers from the NLP and AI communities in the field of deep learning for mathematical reasoning. We study various types of mathematical reasoning problems, such as math word problems, theorem proving, geometry problem solving, math question answering, and other quantitative problems (§2, §A). Additionally, we explore different deep learning architectures for mathematical reasoning, including neural networks" + } + ] + } + ], + "index": 11 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 80, + 761, + 181, + 772 + ], + "type": "page_footnote", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 761, + 181, + 772 + ], + "spans": [ + { + "bbox": [ + 80, + 761, + 181, + 772 + ], + "type": "text", + "content": "* denotes co-senior authors." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 283, + 780, + 312, + 791 + ], + "type": "text", + "content": "14605" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 135, + 795, + 458, + 805 + ], + "type": "footer", + "angle": 0, + "lines": [ + { + "bbox": [ + 135, + 795, + 458, + 805 + ], + "spans": [ + { + "bbox": [ + 135, + 795, + 458, + 805 + ], + "type": "text", + "content": "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 215, + 806, + 377, + 817 + ], + "type": "footer", + "angle": 0, + "lines": [ + { + "bbox": [ + 215, + 806, + 377, + 817 + ], + "spans": [ + { + "bbox": [ + 215, + 806, + 377, + 817 + ], + "type": "text", + "content": "Volume 1: Long Papers, pages 14605-14631" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 176, + 818, + 417, + 829 + ], + "type": "footer", + "angle": 0, + "lines": [ + { + "bbox": [ + 176, + 818, + 417, + 829 + ], + "spans": [ + { + "bbox": [ + 176, + 818, + 417, + 829 + ], + "type": "text", + "content": "July 9-14, 2023 ©2023 Association for Computational Linguistics" + } + ] + } + ], + "index": 16 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 0 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 70, + 63, + 526, + 371 + ], + "blocks": [ + { + "bbox": [ + 70, + 63, + 526, + 371 + ], + "lines": [ + { + "bbox": [ + 70, + 63, + 526, + 371 + ], + "spans": [ + { + "bbox": [ + 70, + 63, + 526, + 371 + ], + "type": "image", + "image_path": "6e2b70a2a59a76734fc50ce31623817c3e58c10156335d7f1cd45aa24bd3f124.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 67, + 381, + 525, + 407 + ], + "lines": [ + { + "bbox": [ + 67, + 381, + 525, + 407 + ], + "spans": [ + { + "bbox": [ + 67, + 381, + 525, + 407 + ], + "type": "text", + "content": "Figure 1: Taxonomy of deep learning for mathematical reasoning. The associated tasks are elaborated in §2, with a comprehensive dataset list found in §A. Deep learning methods are further discussed in §3, §4, and §5." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "spans": [ + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "inline_equation", + "content": "(\\S 3)" + }, + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "text", + "content": ", pre-trained language models " + }, + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "inline_equation", + "content": "(\\S 4)" + }, + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "text", + "content": ", and recent in-context learning for large language models " + }, + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "inline_equation", + "content": "(\\S 5)" + }, + { + "bbox": [ + 67, + 418, + 290, + 445 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 446, + 291, + 581 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 446, + 291, + 581 + ], + "spans": [ + { + "bbox": [ + 67, + 446, + 291, + 581 + ], + "type": "text", + "content": "We also analyze existing benchmarks and find that there is less focus on multi-modal and low-resource settings (§6.1). Our evidence-based studies suggest that current numeracy representations are insufficient and deep learning methods are inconsistent for mathematical reasoning (§6.2). Following this, we suggest future research directions related to generalization and robustness, trustworthy reasoning, learning from feedback, and multimodal mathematical reasoning (§7)." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 590, + 248, + 604 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 590, + 248, + 604 + ], + "spans": [ + { + "bbox": [ + 67, + 590, + 248, + 604 + ], + "type": "text", + "content": "2 Mathematical Reasoning Tasks" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 611, + 290, + 665 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 611, + 290, + 665 + ], + "spans": [ + { + "bbox": [ + 67, + 611, + 290, + 665 + ], + "type": "text", + "content": "In this section, we briefly introduce different tasks for mathematical reasoning. A detailed summary and discussion of commonly used datasets can be found in Table 7 and Appendix A." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 665, + 291, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 665, + 291, + 773 + ], + "spans": [ + { + "bbox": [ + 67, + 665, + 291, + 773 + ], + "type": "text", + "content": "Math Word Problem Solving. Developing algorithms to automatically solve math word problems (MwPs) has been of interest to NLP researchers for decades (Feigenbaum et al., 1963; Bobrow, 1964). An example of a MwP is shown in Table 1. A question involves four basic arithmetic operations with single or multiple operation steps. The challenge posed by MwPs lies in the need for language com" + } + ] + } + ], + "index": 6 + }, + { + "type": "table", + "bbox": [ + 314, + 416, + 514, + 471 + ], + "blocks": [ + { + "bbox": [ + 314, + 416, + 514, + 471 + ], + "lines": [ + { + "bbox": [ + 314, + 416, + 514, + 471 + ], + "spans": [ + { + "bbox": [ + 314, + 416, + 514, + 471 + ], + "type": "table", + "html": "
Question: Bod has 2 apples and David has 5 apples. \nHow many apples do they have in total?
Rationale: x = 2 + 5
Solution: 7
", + "image_path": "2411305be161791887c782b924efbe5aaaa4835669c2e41980cb17b32e2f9e3c.jpg" + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "table_body" + } + ], + "index": 7 + }, + { + "bbox": [ + 334, + 475, + 492, + 487 + ], + "lines": [ + { + "bbox": [ + 334, + 475, + 492, + 487 + ], + "spans": [ + { + "bbox": [ + 334, + 475, + 492, + 487 + ], + "type": "text", + "content": "Table 1: A typical math word problem." + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 302, + 502, + 524, + 528 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 502, + 524, + 528 + ], + "spans": [ + { + "bbox": [ + 302, + 502, + 524, + 528 + ], + "type": "text", + "content": "prehension, semantic parsing, and the application of multiple mathematical reasoning skills." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 529, + 525, + 650 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 529, + 525, + 650 + ], + "spans": [ + { + "bbox": [ + 302, + 529, + 525, + 650 + ], + "type": "text", + "content": "Theorem Proving. Automating theorem proving is a long-standing challenge in AI (Newell et al., 1957; Feigenbaum et al., 1963). The problem is to demonstrate the truth of a mathematical claim (a theorem) through a sequence of logical arguments (a proof). Theorem proving tests various skills, such as choosing effective multi-step strategies, using background knowledge, and performing symbolic manipulations." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 651, + 526, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 651, + 526, + 772 + ], + "spans": [ + { + "bbox": [ + 302, + 651, + 526, + 772 + ], + "type": "text", + "content": "Geometry Problem Solving. Automated geometry problem solving (GPS) is also a long-standing mathematical reasoning task (Gelernter et al., 1960; Wen-Tsun, 1986). As shown in Figure 2, a geometry problem consists of a textual description and a diagram. The multimodal inputs describe the entities, attributes, and relationships of geometric elements, and the goal is to find the numeric solution to an unknown variable." + } + ] + } + ], + "index": 11 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "text", + "content": "14606" + } + ] + } + ], + "index": 12 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 1 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 69, + 68, + 157, + 124 + ], + "blocks": [ + { + "bbox": [ + 69, + 68, + 157, + 124 + ], + "lines": [ + { + "bbox": [ + 69, + 68, + 157, + 124 + ], + "spans": [ + { + "bbox": [ + 69, + 68, + 157, + 124 + ], + "type": "image", + "image_path": "0235c4ea9fe87050bf28bc3e7fcc638381ae97601ae9fea0c1706bd3a2bcbe5b.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 86, + 133, + 271, + 147 + ], + "lines": [ + { + "bbox": [ + 86, + 133, + 271, + 147 + ], + "spans": [ + { + "bbox": [ + 86, + 133, + 271, + 147 + ], + "type": "text", + "content": "Figure 2: An example of geometry problems." + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "spans": [ + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "text", + "content": "Question: In triangle " + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "inline_equation", + "content": "ABC, AD = 3" + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "inline_equation", + "content": "BD = 14" + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "text", + "content": ". Find " + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "inline_equation", + "content": "CD" + }, + { + "bbox": [ + 159, + 74, + 293, + 93 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 159, + 95, + 295, + 106 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 159, + 95, + 295, + 106 + ], + "spans": [ + { + "bbox": [ + 159, + 95, + 295, + 106 + ], + "type": "text", + "content": "Choices: (A) 6.0 (B) 6.5 (C) 7.0 (D) 8.5" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 159, + 109, + 217, + 119 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 159, + 109, + 217, + 119 + ], + "spans": [ + { + "bbox": [ + 159, + 109, + 217, + 119 + ], + "type": "text", + "content": "Answer: (B) 6.5" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 163, + 292, + 272 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 163, + 292, + 272 + ], + "spans": [ + { + "bbox": [ + 67, + 163, + 292, + 272 + ], + "type": "text", + "content": "Math Question Answering. There is a wide range of question answering (QA) benchmarks that center around mathematical reasoning, which we refer to as math question answering (MathQA). For example, DROP (Dua et al., 2019) is a MathQA dataset that requires discrete reasoning to answer questions such as \"Which kicker kicked the most field goals?\" over the content of paragraphs." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 288, + 269, + 317 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 288, + 269, + 317 + ], + "spans": [ + { + "bbox": [ + 67, + 288, + 269, + 317 + ], + "type": "text", + "content": "3 Neural Networks for Mathematical Reasoning" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 329, + 291, + 437 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 329, + 291, + 437 + ], + "spans": [ + { + "bbox": [ + 67, + 329, + 291, + 437 + ], + "type": "text", + "content": "Neural networks have become a popular tool in the field of mathematical reasoning, mirroring their success in NLP. In recent years, a number of different neural network architectures have been proposed for mathematical reasoning tasks, including Seq2Seq-based networks, graph-based networks, and attention-based networks. These methods are outlined in more detail in Table 8 in the Appendix." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 67, + 454, + 257, + 467 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 454, + 257, + 467 + ], + "spans": [ + { + "bbox": [ + 67, + 454, + 257, + 467 + ], + "type": "text", + "content": "3.1 Seq2Seq-based Networks for Math" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 67, + 476, + 291, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 476, + 291, + 772 + ], + "spans": [ + { + "bbox": [ + 67, + 476, + 291, + 772 + ], + "type": "text", + "content": "Sequence-to-sequence (Seq2Seq) (Sutskever et al., 2014) neural networks have been successfully applied to mathematical reasoning tasks, such as math word problem solving (Wang et al., 2017), theorem proving (Yang and Deng, 2019), geometry problem solving (Robaidek et al., 2018), and math question answering (Tafjord et al., 2019). A Seq2Seq model uses an encoder-decoder architecture and usually formalizes mathematical reasoning as a sequence generation task. The basic idea behind this approach is to map an input sequence (e.g. a mathematical problem) to an output sequence (e.g. an equation, program, and proof). Common encoders and decoders include Long Short Term Memory network (LSTM) (Hochreiter and Schmidhuber, 1997), Gated Recurrent Unit (GRU) (Cho et al., 2014), and their bidirectional variants: BiLSTM and BiGRU. A large amount of work has shown the performance advantage of Seq2Seq models over previous statistical learning approaches (Ling et al., 2017; Wang et al., 2018a; Huang et al., 2018; Wang et al., 2019; Li et al., 2019)." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 71, + 485, + 84 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 485, + 84 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 485, + 84 + ], + "type": "text", + "content": "3.2 Graph-based Networks for Math" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 95, + 527, + 433 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 95, + 527, + 433 + ], + "spans": [ + { + "bbox": [ + 302, + 95, + 527, + 433 + ], + "type": "text", + "content": "Seq2Seq approaches show their advantages of generating mathematical expressions without relying on hand-crafted features. It is noteworthy that mathematical expressions can be represented as tree-based structures, such as abstract syntax trees (ASTs) and graph-based structures, which capture the structural information in the expressions. However, Seq2Seq methods do not explicitly this important information. To address this limitation, graph-based neural networks have been developed to explicitly model the structure within expressions. Sequence-to-tree (Seq2Tree) models explicitly model the tree structure when encoding the output sequences (Xie and Sun, 2019; Wu et al., 2020; Zaporojets et al., 2021; Qin et al., 2021). For example, Liu et al. (2019a) devise a Seq2Tree model to better use information from an equation's AST. Seq2DAG (Cao et al., 2021), instead, applies a sequence-to-graph (Seq2Graph) framework when generating the equations since the graph decoder is able to extract complex relationships among multiple variables. The graph-based information can also be embedded when encoding the input mathematical sequences (Zhang et al., 2020b; Shen and Jin, 2020; Li et al., 2020b; Wu et al., 2021a)." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 452, + 498, + 465 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 452, + 498, + 465 + ], + "spans": [ + { + "bbox": [ + 302, + 452, + 498, + 465 + ], + "type": "text", + "content": "3.3 Attention-based Networks for Math" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 476, + 526, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 476, + 526, + 772 + ], + "spans": [ + { + "bbox": [ + 302, + 476, + 526, + 772 + ], + "type": "text", + "content": "The attention mechanism has been successfully applied to NLP (Bahdanau et al., 2015) and vision problems (Xu et al., 2015; Woo et al., 2018), taking into account the hidden vectors of the inputs during the decoding processing. Recently, researchers have been exploring its usefulness in mathematical reasoning tasks, as it can be used to identify the most important relationships between mathematical concepts. For instance, MATH-EN (Wang et al., 2018a) is a math word problem solver which benefits from long-distance dependency information learned by self-attention. Attention-based methods have also been applied to other mathematical reasoning tasks such as geometry problems solving (Robaidek et al., 2018; Chen et al., 2021a) and theorem proving (Yang and Deng, 2019). Various attention mechanisms have been studied to extract better representations, such as Group-ATT (Li et al., 2019) which uses different multi-head attention to extract various types of MWP features, and graph attention which is applied to extract knowledge-aware information in (Wu et al., 2020)." + } + ] + } + ], + "index": 13 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "text", + "content": "14607" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 2 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 251, + 84 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 251, + 84 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 251, + 84 + ], + "type": "text", + "content": "3.4 Other Neural Networks for Math" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 89, + 291, + 264 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 89, + 291, + 264 + ], + "spans": [ + { + "bbox": [ + 67, + 89, + 291, + 264 + ], + "type": "text", + "content": "Deep learning approaches to mathematical reasoning tasks can also make use of other neural networks, such as convolutional neural networks (CNN) and multimodal networks. Some work encodes the input text using a convolutional neural network architecture, giving the model the ability to capture long-term relationships between symbols in the input (Gehring et al., 2017; Wang et al., 2018a,a; Robaidek et al., 2018; Alemi et al., 2016; Loos et al., 2017). For example, the first application of deep neural networks for theorem proving is proposed in (Alemi et al., 2016), which relies on convolutional networks for premise selection." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 266, + 291, + 427 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 266, + 291, + 427 + ], + "spans": [ + { + "bbox": [ + 69, + 266, + 291, + 427 + ], + "type": "text", + "content": "Multimodal mathematical reasoning tasks, such as geometry problem solving and diagram-based mathematical reasoning, are formalized as visual question answer (VQA) problems (Kafle et al., 2018; Chen et al., 2021a; Lu et al., 2021b). In this domain, visual inputs are encoded using ResNet (He et al., 2016) or Faster-RCNN (Ren et al., 2015), while textual representations are obtained via GRU or LTSM. Subsequently, the joint representation is learned using multimodal fusion models, such as BAN (Kim et al., 2018), FiLM (Perez et al., 2018), and DAFA (Gao et al., 2019)." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 428, + 291, + 618 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 428, + 291, + 618 + ], + "spans": [ + { + "bbox": [ + 67, + 428, + 291, + 618 + ], + "type": "text", + "content": "Other deep neural network structures can also be used in mathematical reasoning. A Graph Neural Network (GNN) is employed for geometry problem parsing in Zhang et al. (2022), taking advantage of its success in spatial reasoning. WaveNet has been applied to theorem proving (Loos et al., 2017; Bansal et al., 2019), due to its ability to address longitudinal time-series data. Furthermore, Transformers are found to outperform GRU in generating mathematical equations in DDT (Meng and Rumshisky, 2019). Finally, MathDQN (Wang et al., 2018b) is the first work to explore reinforcement learning for math word problem solving, taking advantage of its strong search capabilities." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 629, + 260, + 657 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 629, + 260, + 657 + ], + "spans": [ + { + "bbox": [ + 67, + 629, + 260, + 657 + ], + "type": "text", + "content": "4 Pre-trained Language Models for Mathematical Reasoning" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 665, + 291, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 665, + 291, + 773 + ], + "spans": [ + { + "bbox": [ + 67, + 665, + 291, + 773 + ], + "type": "text", + "content": "Pre-trained language models (Devlin et al., 2019; Radford et al., 2020; Brown et al., 2020) have demonstrated remarkable performance gains on a wide range of NLP tasks. By pre-training on a large corpus of text, the models learn valuable world knowledge (Guu et al., 2020), which could be applied to downstream tasks. Similar ideas can be applied to math-related problems, and previous" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 71, + 526, + 138 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 526, + 138 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 526, + 138 + ], + "type": "text", + "content": "work has shown the promising performance of pretrained language models in answering math word problems (Kim et al., 2020), assisting with theorem proving (Wu et al., 2022b), as well as solving other mathematical tasks (Charton, 2022)." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 141, + 526, + 397 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 141, + 526, + 397 + ], + "spans": [ + { + "bbox": [ + 302, + 141, + 526, + 397 + ], + "type": "text", + "content": "However, though large language models excel in modeling natural language, there are several challenges to using them for mathematical reasoning. First, pre-trained language models are not specifically trained on mathematical data. This likely contributes to them being less proficient in math-related tasks compared to natural language tasks. There is also less mathematical or scientific data available for large-scale pre-training compared to text data. Second, the size of pre-trained models continues to grow, making it expensive to train the entire model from scratch for specific downstream tasks. Additionally, downstream tasks may deal with different input formats or modalities, such as structured tables (Zhao et al., 2022) or diagrams (Lu et al., 2021b). To address these challenges, researchers have to adjust pre-trained models by finetuning them on downstream tasks or adapting the neural architectures." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 412, + 495, + 426 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 412, + 495, + 426 + ], + "spans": [ + { + "bbox": [ + 302, + 412, + 495, + 426 + ], + "type": "text", + "content": "4.1 Self-Supervised Learning for Math" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 433, + 525, + 514 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 433, + 525, + 514 + ], + "spans": [ + { + "bbox": [ + 302, + 433, + 525, + 514 + ], + "type": "text", + "content": "Self-supervised learning is a machine learning approach in which an algorithm learns to perform a task without being explicitly provided with labeled training data. Table 2 provides a list of language models pre-trained with self-supervised tasks for mathematical reasoning." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 516, + 526, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 516, + 526, + 773 + ], + "spans": [ + { + "bbox": [ + 302, + 516, + 526, + 773 + ], + "type": "text", + "content": "Model scale. There is a clear trend that pre-trained language models have become increasingly larger in the past few years (Devlin et al., 2019; Lewis et al., 2020; Raffel et al., 2020; Radford et al., 2020; Brown et al., 2020). A recent study (Liang et al., 2022a) shows that model scale within a model family reliably predicts model accuracy. The study also mentions an interesting thresholding effect: \"all models that win head-to-head model comparisons for accuracy at a rate well above chance are at least 50B parameters\". A similar size-growing trend can be observed in the field of mathematical reasoning with pre-trained language models. For example, MWP-BERT (Liang et al., 2022b) uses a backbone of BERT (110M) (Devlin et al., 2019) and RoBERTa (123M) (Liu et al., 2019b) for Math Word Problems. Most recently, Minerva (Lewkowycz et al., 2022), which is based on the PaLM (Chowdhery et al., 2022) pre-trained" + } + ] + } + ], + "index": 10 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "type": "text", + "content": "14608" + } + ] + } + ], + "index": 11 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 3 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 73, + 68, + 520, + 222 + ], + "blocks": [ + { + "bbox": [ + 73, + 68, + 520, + 222 + ], + "lines": [ + { + "bbox": [ + 73, + 68, + 520, + 222 + ], + "spans": [ + { + "bbox": [ + 73, + 68, + 520, + 222 + ], + "type": "table", + "html": "
PaperBackboneSizeCorpusPre-training task
GPT-f (Polu and Sutskever, 2020)Transformer (2017)774MMathCausal language modeling
LISA (Jiang et al., 2021)Transformer (2017)163MMathCausal language modeling
MATH-PLM (Hendrycks et al., 2021b)GPT-2 (2020)1.5BMathCausal language modeling
MWP-BERT (Liang et al., 2022b)RoBERTa (2019b)123MMath8 numeracy augmented tasks
TaPEx (Liu et al., 2022b)BART (2020)406MSQLQuery result generation
HTTPS (Lample et al., 2022)Transformer (2017)600MMathMasked Seq2Seq modeling
Thor (Jiang et al., 2022b)Transformer (2017)700MGithub, arXivCausal language modeling
PACT (Han et al., 2022)Transformer (2017)837MMathMasked/Causal language modeling
Minerva (Lewkowycz et al., 2022)PaLM (2022)540BScience & MathCausal language modeling
GenBERT (Geva et al., 2020)BERT (2019)110MNumber, TextMasked/Causal language modeling
NF-NSM (Feng et al., 2021)RoBERTa (2019b)110MNumberNumber prediction
LIME (Wu et al., 2021d)Transformer (2017)11BMathCausal language modeling
Set (Wu et al., 2022c)T5 (2020)60MMathUnique token generation
", + "image_path": "6c3080f6af028a03867cfefcfb383efc17179d2ce22204e2c87538460b0ccc8a.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 252, + 291, + 266 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 252, + 291, + 266 + ], + "spans": [ + { + "bbox": [ + 67, + 252, + 291, + 266 + ], + "type": "text", + "content": "language model, has a size up to 540B parameters." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 66, + 269, + 291, + 703 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 66, + 269, + 291, + 703 + ], + "spans": [ + { + "bbox": [ + 66, + 269, + 291, + 703 + ], + "type": "text", + "content": "Pre-training corpus. There are generally two types of pre-training corpus for mathematical language models. (i) Curated datasets from openly accessible sources. For example, Hendrycks et al. (2021b) present the first large-scale mathematics pre-training dataset with step-by-step solutions in natural language and LATEX, called the Auxiliary Mathematics Problems and Solutions (AMPS). AMPS consists of Khan Academy and Mathematica data. Minerva (Lewkowycz et al., 2022) collects a high-quality dataset containing scientific and mathematical data, which contains 38.5B tokens from webpages filtered for mathematical content and from papers submitted to the arXiv preprint server. Thor (Jiang et al., 2022b) pre-trains a language model on the GitHub + arXiv subsets of The Pile (Gao et al., 2020). (ii) Synthetic datasets based on templates or interaction with engines. Recent work (Wu et al., 2021d; Krishna et al., 2021; Ri and Tsuruoka, 2022; Anderson and Farrell, 2022; Wu et al., 2022c) shows that pre-training on data that is fully synthetically generated—synthetic pre-training can actually provide substantial gains. Representative work includes TaPEX (Liu et al., 2022b), which obtains a pre-training corpus by automatically synthesizing executable SQL queries and their execution outputs. LISA (Jiang et al., 2021) extracts lemmas and theorems by interacting with the Isabelle standard library and the Archive of Formal Proofs. GenBERT (Geva et al., 2020) generates numerical and textual pre-training datasets based on manually crafted and extracted templates." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 705, + 291, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 705, + 291, + 774 + ], + "spans": [ + { + "bbox": [ + 67, + 705, + 291, + 774 + ], + "type": "text", + "content": "Pre-training tasks. General pre-training language models have two typical self-supervised learning tasks: (i) Masked Language Modeling (MLM), where it randomly masks a portion of words in each sequence to predict the outcome; (ii) Causal Lan" + } + ] + } + ], + "index": 4 + }, + { + "type": "table", + "bbox": [ + 304, + 249, + 526, + 485 + ], + "blocks": [ + { + "bbox": [ + 130, + 224, + 461, + 238 + ], + "lines": [ + { + "bbox": [ + 130, + 224, + 461, + 238 + ], + "spans": [ + { + "bbox": [ + 130, + 224, + 461, + 238 + ], + "type": "text", + "content": "Table 2: Comparison of pre-training language models for mathematical reasoning." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 304, + 249, + 526, + 485 + ], + "lines": [ + { + "bbox": [ + 304, + 249, + 526, + 485 + ], + "spans": [ + { + "bbox": [ + 304, + 249, + 526, + 485 + ], + "type": "table", + "html": "
PaperBackboneTask
EPT (2020)ALBERT (2019)MWP
Generate & Rank (2021)BART (2020)MWP
RPKHS (2021b)RoBERTa (2019b)MWP
PatchTRM (2021b)ResNet+BERT (2019)MWP
GSM8K-PLM (2021)GPT-3 (2020)MWP
BERT-TD+CL (2022b)BERT (2019)MWP
DeductReasoner (2022)RoBERTa (2019b)MWP
Self-Sampling (2023)GPT-Neo (2020)MWP
Bhaskara (2022a)GPT-Neo (2020)MWP
miniF2F-PLM (2022)GPT-f (2020)TP
NaturalProver (2022a)GPT-3 (2020)TP
Inter-GPS (2021a)BART (2020)GPS
UniGeo (2022a)VL-T5 (2021)GPS
DPE-NGS (2022)RoBERTa (2019b)GPS
Aristo (2020)RoBERTa (2019b)MathQA
FinQANet (2021c)RoBERTa (2019b)MathQA
TAGOP (2021)RoBERTa (2019b)MathQA
MT2Net (2022)RoBERTa (2019b)MathQA
Scratchpad (2021)Transformer (2017)Mixed
LAMT (2022)Transformer (2017)Mixed
", + "image_path": "bc9543c12f6b6cf20ab3d39b917785c53743f15f98108db0a40b09e70d68c25b.jpg" + } + ] + } + ], + "index": 5, + "angle": 0, + "type": "table_body" + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 488, + 525, + 513 + ], + "lines": [ + { + "bbox": [ + 302, + 488, + 525, + 513 + ], + "spans": [ + { + "bbox": [ + 302, + 488, + 525, + 513 + ], + "type": "text", + "content": "Table 3: Finetuned pre-trained language models for downstream mathematical reasoning tasks." + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 301, + 528, + 525, + 623 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 301, + 528, + 525, + 623 + ], + "spans": [ + { + "bbox": [ + 301, + 528, + 525, + 623 + ], + "type": "text", + "content": "guage Modeling (CLM), where the model is trained to predict the next token in a sequence of tokens. Following the same paradigm, researchers pre-train language models with MLM and CLM tasks on mathematical or scientific corpora for downstream tasks (Polu and Sutskever, 2020; Hendrycks et al., 2021b; Han et al., 2022; Jiang et al., 2022b)." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 301, + 624, + 526, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 301, + 624, + 526, + 773 + ], + "spans": [ + { + "bbox": [ + 301, + 624, + 526, + 773 + ], + "type": "text", + "content": "There is also recent work that designs customized tasks to inject mathematical reasoning capabilities into language models. For instance, Liang et al. (2022b) pre-train language models with a suite of 8 numeracy-augmented tasks with consideration of reasoning logic and numerical properties. LIME (Wu et al., 2021d) proposes synthetic pretraining tasks to learn three reasoning primitives: deduction, induction, and abduction before learning more complex reasoning skills, which also be regarded as a form of curriculum learning." + } + ] + } + ], + "index": 8 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "text", + "content": "14609" + } + ] + } + ], + "index": 9 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 4 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 257, + 84 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 257, + 84 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 257, + 84 + ], + "type": "text", + "content": "4.2 Task-specific Fine-tuning for Math" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 89, + 291, + 292 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 89, + 291, + 292 + ], + "spans": [ + { + "bbox": [ + 67, + 89, + 291, + 292 + ], + "type": "text", + "content": "Task-specific fine-tuning is a technique to improve the performance of a pre-trained language model on a specific task. This is also a common practice when there is not enough data for training the large models from scratch. As shown in Table 3, existing work fine-tunes pre-trained language models on a variety of downstream tasks, such as math word problems (Kim et al., 2020; Shen et al., 2021), MathQA (Zhao et al., 2022), geometry problem solving (Lu et al., 2021a), linear algebra (Charton, 2022), and theorem proving (Welleck et al., 2022a). Apart from fine-tuning the model parameters, some work also uses pre-trained language models as encoders and ensembles them with other modules for downstream tasks (Lu et al., 2021b)." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 303, + 285, + 332 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 303, + 285, + 332 + ], + "spans": [ + { + "bbox": [ + 67, + 303, + 285, + 332 + ], + "type": "text", + "content": "5 In-context Learning for Mathematical Reasoning" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 339, + 291, + 528 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 339, + 291, + 528 + ], + "spans": [ + { + "bbox": [ + 67, + 339, + 291, + 528 + ], + "type": "text", + "content": "Large language models (LLMs), such as GPT-3 (Brown et al., 2020), have recently revolutionized the field of natural language processing (NLP), especially on account of their powerful few-shot in-context learning capabilities (Brown et al., 2020). In-context Learning (ICL) enables LLMs to perform target tasks by providing some task examples as conditions at inference time, without updating model parameters (Radford et al., 2020; Brown et al., 2020). ICL allows users to quickly build models for new use cases without worrying about fine-tuning and storing a large amount of new parameters for each task, so it is widely used in few-shot settings nowadays (Min et al., 2022)." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 529, + 291, + 690 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 529, + 291, + 690 + ], + "spans": [ + { + "bbox": [ + 67, + 529, + 291, + 690 + ], + "type": "text", + "content": "An in-context example typically contains an input-output pair with some prompt words, e.g., Please select the largest number from the list. Input: [2, 4, 1, 5, 8]. Output: 8, and few-shot works by giving multiple examples, and then a final input example, where the model is expected to predict the output. However, such standard few-shot prompts, in which the LLM is given in-context examples of input-output pairs in front of test-time examples, have not yet proved sufficient to achieve high performance on challenging tasks such as mathematical reasoning (Rae et al., 2021)." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 692, + 291, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 692, + 291, + 772 + ], + "spans": [ + { + "bbox": [ + 67, + 692, + 291, + 772 + ], + "type": "text", + "content": "Chain-of-thought prompting (CoT) (Wei et al., 2022) leverages intermediate natural language rationales as prompts to enable LLMs to first generate reasoning chains and then predict an answer for an input question. For example, a CoT prompt for solving the math word problem could be" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 324, + 71, + 504, + 125 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 324, + 71, + 504, + 125 + ], + "spans": [ + { + "bbox": [ + 324, + 71, + 504, + 125 + ], + "type": "text", + "content": "Question: Roger has 5 tennis balls. He buys 2 more cans of tennis balls. Each can has 3 tennis balls. Then, how many tennis balls does Roger have now?" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "spans": [ + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "type": "text", + "content": "Answer: Roger started with 5 balls. 2 cans of 3 tennis balls each are 6 tennis balls. " + }, + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "type": "inline_equation", + "content": "5 + 6 = 11" + }, + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "type": "text", + "content": ". The answer is " + }, + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "type": "inline_equation", + "content": "\\underline{11}" + }, + { + "bbox": [ + 324, + 126, + 504, + 166 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 177, + 527, + 285 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 177, + 527, + 285 + ], + "spans": [ + { + "bbox": [ + 302, + 177, + 527, + 285 + ], + "type": "text", + "content": "Apart from Kojima et al. (2022) showing that LLMs are decent zero-shot reasoners when given the \"Let's think step by step!\" prompt, most of the recent work has focused on how to improve chain-of-thought reasoning under the few-shot setting. This work is mainly divided into two parts, (i) selecting better in-context examples and (ii) creating better reasoning chains." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 295, + 468, + 308 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 295, + 468, + 308 + ], + "spans": [ + { + "bbox": [ + 302, + 295, + 468, + 308 + ], + "type": "text", + "content": "5.1 In-context Example Selection" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 312, + 527, + 597 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 312, + 527, + 597 + ], + "spans": [ + { + "bbox": [ + 302, + 312, + 527, + 597 + ], + "type": "text", + "content": "Early chain-of-thought work randomly or heuristically selects in-context examples. However, recent studies have shown that this type of few-shot learning can be highly unstable across different selections of in-context examples (Rubin et al., 2022; Liu et al., 2022a). Therefore, which in-context reasoning examples make the most effective prompts is still an unknown problem in the literature. To address the limitation, recent work has investigated various methods to optimize the in-context examples selection process (Rubin et al., 2022; Zhang et al., 2023; Lu et al., 2022b; Yu et al., 2023; Fu et al., 2023). For example, Rubin et al. (2022) attempt to address this issue by retrieving semantically similar examples. In addition, Fu et al. (2023) propose complexity-based prompting, which chooses examples with complex reasoning chains, i.e., chains with more reasoning steps, as the prompt. PromptPG (Lu et al., 2022b) learns to select optimal in-context examples via reinforcement learning (RL) from a candidate pool." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 607, + 478, + 619 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 607, + 478, + 619 + ], + "spans": [ + { + "bbox": [ + 302, + 607, + 478, + 619 + ], + "type": "text", + "content": "5.2 High-quality Reasoning Chains" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 624, + 527, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 624, + 527, + 772 + ], + "spans": [ + { + "bbox": [ + 302, + 624, + 527, + 772 + ], + "type": "text", + "content": "Early chain of thought work (e.g., Wei et al. (2022)) mainly relies on a single human-annotated reasoning chain as a prompt. However, manually creating reasoning chains has two disadvantages. First, as tasks become more complex, current models may not be sufficient to learn to perform all necessary reasoning steps and cannot easily generalize to different tasks. Second, a single decoding process is vulnerable to incorrect inference steps, leading to an incorrect prediction as the final answer. To address this limitation, recent studies mainly fo" + } + ] + } + ], + "index": 12 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "type": "text", + "content": "14610" + } + ] + } + ], + "index": 13 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 5 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 71, + 60, + 520, + 176 + ], + "blocks": [ + { + "bbox": [ + 71, + 60, + 520, + 176 + ], + "lines": [ + { + "bbox": [ + 71, + 60, + 520, + 176 + ], + "spans": [ + { + "bbox": [ + 71, + 60, + 520, + 176 + ], + "type": "table", + "html": "
ModelsEngine (best performed)ICL sourceRationale typeRationale sourcePost method
Few-shot-CoT (Wei et al., 2022)PaLM (540B)RandomLanguageHand-crafted-
Self-Consistency-CoT (Wang et al., 2023)Codex (175B)RandomLanguageHand-craftedSelf-consistency
Least-to-most CoT (Zhou et al., 2023)Codex (175B)RandomLanguageHand-crafted-
PromptPG-CoT (Lu et al., 2022b)GPT-3 (175B)RLLanguageHand-crafted-
Retrieval-CoT (Zhang et al., 2023)GPT-3 (175B)RetrivalLanguageAuto-generated-
Auto-CoT (Zhang et al., 2023)Codex (175B)ClusteringLanguageAuto-generated-
Complexity-CoT (Fu et al., 2023)GPT-3 (175B)ComplexityLanguageHand-craftedSelf-consistency
Few-shot-PoT (Chen et al., 2022b)GPT-3 (175B)RandomCodeHand-crafted-
", + "image_path": "01c7af74d0cb2796cbba7978e2ae8b8f02ac77dac6fa6c53194d8a3e3c4bfc32.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 179, + 525, + 204 + ], + "lines": [ + { + "bbox": [ + 67, + 179, + 525, + 204 + ], + "spans": [ + { + "bbox": [ + 67, + 179, + 525, + 204 + ], + "type": "text", + "content": "Table 4: In-context learning with large language models for mathematical reasoning. For GPT-3, all papers use the text-davinci-002 version; for Codex, all papers use the code-davinci-002. RL is short for reinforcement learning." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 67, + 216, + 290, + 297 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 216, + 290, + 297 + ], + "spans": [ + { + "bbox": [ + 67, + 216, + 290, + 297 + ], + "type": "text", + "content": "cus on two aspects, (i) hand-crafting more complex demonstrations, which we refer to as process-based approaches (Zhou et al., 2023; Chen et al., 2022b), (ii) leveraging ensemble-like methods, which we refer to as outcome-based approaches (Wang et al., 2023; Li et al., 2022a)." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 305, + 291, + 602 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 305, + 291, + 602 + ], + "spans": [ + { + "bbox": [ + 69, + 305, + 291, + 602 + ], + "type": "text", + "content": "Process-based approaches aim to improve the chain-of-thought reasoning quality, especially for complex reasoning tasks. In least-to-most prompting (Zhou et al., 2023), the problem-solving process is implemented through two-stage prompting: (i) reducing a complex problem into a list of subproblems; (ii) solving these sub-problems sequentially, so that solving a given sub-problem is facilitated by the answers to previously solved subproblems. Similarly, Khot et al. (2022) leverage diverse decomposition structures and use different prompts to answer each sub-question. Apart from these multi-step reasoning methods, Chen et al. (2022b); Gao et al. (2022) propose program-of-thoughts (PoT), an alternative solution that uses large language models to express the reasoning process as a program. The computation is then relegated to an external computer, which executes the generated programs to derive the answer. A more recent work, Chameleon (Lu et al., 2023), integrates different tools to enhance the abilities of LLMs for compositional reasoning." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 611, + 291, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 611, + 291, + 773 + ], + "spans": [ + { + "bbox": [ + 67, + 611, + 291, + 773 + ], + "type": "text", + "content": "Outcome-based approaches acknowledge the potential incorrectness of an individual reasoning path, and instead use multiple reasoning paths (Wang et al., 2023; Li et al., 2022a). Self-consistency (Wang et al., 2023) generates a set of reasoning paths by sampling from the language model, and marginalizes out the reasoning paths by choosing the most common answer. In addition to using sampling with a single prompt to produce multiple reasoning paths, Li et al. (2022a) propose to introduce diverse prompts through \"self-teaching\", as a complementary solution to produce" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 302, + 216, + 427, + 229 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 216, + 427, + 229 + ], + "spans": [ + { + "bbox": [ + 302, + 216, + 427, + 229 + ], + "type": "text", + "content": "a higher degree of diversity." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 239, + 448, + 253 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 239, + 448, + 253 + ], + "spans": [ + { + "bbox": [ + 302, + 239, + 448, + 253 + ], + "type": "text", + "content": "6 Discussion and Findings" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 261, + 444, + 274 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 261, + 444, + 274 + ], + "spans": [ + { + "bbox": [ + 302, + 261, + 444, + 274 + ], + "type": "text", + "content": "6.1 Analysis of Benchmarks" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 301, + 279, + 526, + 535 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 301, + 279, + 526, + 535 + ], + "spans": [ + { + "bbox": [ + 301, + 279, + 526, + 535 + ], + "type": "text", + "content": "The multi-modal setting is underexplored but is gaining increasing attention. Most existing benchmarks for mathematical reasoning have targeted the textual-only modality. However, visual elements can provide a rich source of quantitative information, making multi-modal datasets beneficial for reasoning over quantitative relations in natural images (Lu et al., 2022a), abstract diagrams (Lu et al., 2021b), figures (Kahou et al., 2018), and charts (Kafle et al., 2018). Tables, which are commonly found in daily documents and contain hierarchically structured information, have also been the focus of tasks that require quantitative reasoning over textual and tabular context (Chen et al., 2021c; Zhu et al., 2021; Zhao et al., 2022; Lu et al., 2022b). In addition, recent datasets have been developed for mathematical reasoning grounded on conversations (Sun et al., 2019; Zhang et al., 2021; Chen et al., 2022c), as well as reports (Chen et al., 2022c)." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 539, + 526, + 702 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 539, + 526, + 702 + ], + "spans": [ + { + "bbox": [ + 302, + 539, + 526, + 702 + ], + "type": "text", + "content": "Pioneering work is emerging in the exploration of low-resource settings. Despite the creation of various datasets, mathematical reasoning in low-resource settings remains largely under-explored. Pioneering research has developed mathematical reasoning benchmarks for financial (Chen et al., 2021c; Zhu et al., 2021; Zhao et al., 2022) and scientific domains (Lu et al., 2022a). Additionally, there have been attempts to build non-English datasets for Chinese (Wang et al., 2017; Qin et al., 2020; Yu et al., 2021a) and Arabic (Alghamdi et al., 2022) for mathematical reasoning." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 706, + 526, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 706, + 526, + 772 + ], + "spans": [ + { + "bbox": [ + 302, + 706, + 526, + 772 + ], + "type": "text", + "content": "Diverse rationale annotations have been widely explored. Complex reasoning usually involves multiple steps to arrive at the final answer. To bridge this gap, datasets annotated with intermediate rationales such as logic forms (Tafjord et al.," + } + ] + } + ], + "index": 10 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 311, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 311, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 311, + 791 + ], + "type": "text", + "content": "14611" + } + ] + } + ], + "index": 11 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 6 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 68, + 60, + 289, + 122 + ], + "blocks": [ + { + "bbox": [ + 68, + 60, + 289, + 122 + ], + "lines": [ + { + "bbox": [ + 68, + 60, + 289, + 122 + ], + "spans": [ + { + "bbox": [ + 68, + 60, + 289, + 122 + ], + "type": "table", + "html": "
T5 (Large)UnifiedQA (Large)GPT-3 (davinci-002)GPT-3 (davinci-003)
3 balls + 5 balls =X5 balls8 balls8 balls
23 balls + 145 balls =XX58 balls168 balls
23 balls + 1,855 balls =XX2,878 balls2,988 balls
", + "image_path": "53f6906fa363e8cf2565cd7f1d1cd009a08bfc6d5729184de2776a5fa1d416f2.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 148, + 291, + 311 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 148, + 291, + 311 + ], + "spans": [ + { + "bbox": [ + 67, + 148, + 291, + 311 + ], + "type": "text", + "content": "2019; Lu et al., 2021a), programs (Amini et al., 2019; Chen et al., 2021c,a; Cao and Xiao, 2022; Chen et al., 2022a), and reasoning graphs (Zhang et al., 2021) have been proposed to train models for complex reasoning tasks. Python programs are used as reasoning annotations in (Austin et al., 2021; Mishra et al., 2022a) due to their enhanced accessibility and readability. To imitate the reasoning process of a human, a more recent trend is to annotate solutions in natural language (Ling et al., 2017; Cobbe et al., 2021; Lu et al., 2022b; Hendrycks et al., 2021b; Lu et al., 2022a)." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 322, + 262, + 335 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 322, + 262, + 335 + ], + "spans": [ + { + "bbox": [ + 67, + 322, + 262, + 335 + ], + "type": "text", + "content": "6.2 Analysis of Deep Learning Methods" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 340, + 291, + 514 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 340, + 291, + 514 + ], + "spans": [ + { + "bbox": [ + 67, + 340, + 291, + 514 + ], + "type": "text", + "content": "Is the current representation of numeracy sufficient? The standard practice for deep learning techniques is to treat numbers in the same way as words. Early neural network methods create a vocabulary that maps input words and numbers to token IDs, resulting in less frequent numbers being collapsed into an \"UNK\" token. Recent language models use subword tokenization techniques (Wu et al., 2016; Sennrich et al., 2016) to split numbers into atomic tokens. Recent studies have shown that these tokenization approaches are suboptimal (Wallace et al., 2019; Lin et al., 2020; Zhang et al., 2020d; Thawani et al., 2022)." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 516, + 292, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 516, + 292, + 773 + ], + "spans": [ + { + "bbox": [ + 67, + 516, + 292, + 773 + ], + "type": "text", + "content": "Two numbers on the same or close number line could have surface forms with no shared common tokens. For example, a number like 1598 is tokenized as “15” and “98” in GPT-3, while another format like 1, 598 is split as three different tokens: “1”, “;”, and “598”. This lack of consistent representation can make it difficult for deep learning models to effectively process numbers, especially when compared to pure text. The insufficient representations of numbers can lead to out-of-distribution (OOD) problems. Table 5 provides examples of where language models tend to struggle with large numbers. Although increasing model scales could help, even the state-of-the-art large language model GPT-3 performs poorly when reasoning over large numbers. Some recent work suggests that using scientific notation (Zhang et al., 2020d) and digit-level decomposition (Geva et al., 2020) may be helpful in improving numeracy representation, but" + } + ] + } + ], + "index": 5 + }, + { + "type": "table", + "bbox": [ + 305, + 61, + 525, + 213 + ], + "blocks": [ + { + "bbox": [ + 68, + 126, + 289, + 138 + ], + "lines": [ + { + "bbox": [ + 68, + 126, + 289, + 138 + ], + "spans": [ + { + "bbox": [ + 68, + 126, + 289, + 138 + ], + "type": "text", + "content": "Table 5: Language models struggle with large numbers." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 305, + 61, + 525, + 213 + ], + "lines": [ + { + "bbox": [ + 305, + 61, + 525, + 213 + ], + "spans": [ + { + "bbox": [ + 305, + 61, + 525, + 213 + ], + "type": "table", + "html": "
ProblemsGPT-3 (text-davinci-002)
John had 8 balls and he gave 3 to Mary.John has 5 balls.
How many balls does John have now?
John had 3 apples.Mary has 5 balls.
he gave 3 to Mary.How many balls does Mary have now?
John had 8 balls and he gave 3 to Mary.John has more balls.
Who has more balls now?
John had 8 balls and he gave 3 to Mary.No, John has 5 balls now.
Does John have more balls now?
John had 8 balls and he gave 4 to Mary.No, John has 4 balls now.
Does John have more balls now?
John had 8 balls and he gave 4 to Mary.John has more balls.
Who has more balls now?
", + "image_path": "3963c1f6a2607eb55d16f77ecf313004a119705936342c3092a3e6dc792f941c.jpg" + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "table_body" + }, + { + "bbox": [ + 302, + 216, + 525, + 241 + ], + "lines": [ + { + "bbox": [ + 302, + 216, + 525, + 241 + ], + "spans": [ + { + "bbox": [ + 302, + 216, + 525, + 241 + ], + "type": "text", + "content": "Table 6: Examples where large language models are not consistent for mathematical reasoning." + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "table_footnote" + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 248, + 439, + 261 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 248, + 439, + 261 + ], + "spans": [ + { + "bbox": [ + 302, + 248, + 439, + 261 + ], + "type": "text", + "content": "this remains an open problem." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "spans": [ + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "type": "text", + "content": "Are deep learning methods consistent for mathematical reasoning? Recent developments in deep learning have led to impressive results on various mathematical reasoning tasks. The zero-shot-CoT Minerva 540B achieves a score of " + }, + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "type": "inline_equation", + "content": "75.0\\%" + }, + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "type": "text", + "content": " on the MMLU-STEM benchmark (Hendrycks et al., 2021a), which assesses multitask reasoning ability in the fields of science, technology, engineering, and mathematics (STEM) at both high school and college levels. Similarly, few-shot-CoT GPT-3 175B achieves a high accuracy of " + }, + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "type": "inline_equation", + "content": "93.0\\%" + }, + { + "bbox": [ + 302, + 265, + 526, + 454 + ], + "type": "text", + "content": " on the MultiArith task. However, the question remains as to whether these methods are sufficiently advanced to tackle more complex problems." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "spans": [ + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "type": "text", + "content": "There is strong evidence that deep learning methods for mathematical reasoning are not robust and susceptible to adversarial attacks (Lin et al., 2020; Patel et al., 2021; Mishra et al., 2022b,a; Welleck et al., 2022b). The SVAMP (Patel et al., 2021) dataset is a collection of one-unknown arithmetic word problems up to grade 4, with slight word variations from previous datasets. It is surprising that current state-of-the-art (SOTA) methods perform poorly on this dataset, with Graph2Tree achieving only a " + }, + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "type": "inline_equation", + "content": "43.8\\%" + }, + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "type": "text", + "content": " accuracy and zero-shot-CoT GPT-3 (175B) only reaching " + }, + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "type": "inline_equation", + "content": "63.7\\%" + }, + { + "bbox": [ + 302, + 455, + 525, + 698 + ], + "type": "text", + "content": ", which is just above an \"F\" grade. Table 6 also shows the inconsistent performance of the zero-shot GPT-3 model in scenarios with slightly different descriptions, while human performance remains unchanged. This indicates a lack of consistency in the mathematical reasoning ability of SOTA large language models." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 303, + 708, + 392, + 719 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 708, + 392, + 719 + ], + "spans": [ + { + "bbox": [ + 303, + 708, + 392, + 719 + ], + "type": "text", + "content": "7 Future Work" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 729, + 476, + 741 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 729, + 476, + 741 + ], + "spans": [ + { + "bbox": [ + 302, + 729, + 476, + 741 + ], + "type": "text", + "content": "7.1 Generalization and Robustness" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 746, + 525, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 746, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 302, + 746, + 525, + 772 + ], + "type": "text", + "content": "Despite impressive progress, neural models commonly display generalization and robustness fail" + } + ] + } + ], + "index": 13 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "text", + "content": "14612" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 7 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 293, + 192 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 293, + 192 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 293, + 192 + ], + "type": "text", + "content": "ures on reasoning tasks. For example, above we discussed difficulties in generalizing to larger numbers (Table 5) or remaining robust to nearby problems (Table 6), while others identify failures in generalizing to longer problems than those observed in training (e.g., Anil et al. (2022)). One direction is to explore new inference-time (Jung et al., 2022; Mitchell et al., 2022) or fine-tuning (Anil et al., 2022) strategies." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 194, + 292, + 396 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 194, + 292, + 396 + ], + "spans": [ + { + "bbox": [ + 69, + 194, + 292, + 396 + ], + "type": "text", + "content": "Another aspect of generalization relates to the role of memorization. For example, is the ability to produce a complex solution dependent on seeing many similar solutions during training, or even on memorizing the solution? Term frequency in the pretraining corpus is known to impact accuracy in simple arithmetic tasks (Razeghi et al., 2022) or factual question answering (Kandpal et al., 2022). On the other hand, Lewkowycz et al. (2022) did not find evidence of memorization in complex outputs, yet their training set and model are not available for inspection. Gaining a full understanding of these factors for complex problems and outputs (e.g., multi-step solutions or proofs) requires more analysis, as well as accessible datasets and models." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 405, + 206, + 417 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 405, + 206, + 417 + ], + "spans": [ + { + "bbox": [ + 67, + 405, + 206, + 417 + ], + "type": "text", + "content": "7.2 Trustworthy Reasoning" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 423, + 291, + 679 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 423, + 291, + 679 + ], + "spans": [ + { + "bbox": [ + 67, + 423, + 291, + 679 + ], + "type": "text", + "content": "Recent advances in language models have demonstrated their powerful capabilities for mathematical reasoning. However, due to the potential for generating ungrounded answers (Nakano et al., 2021), users can't always trust the predicted outcomes or have to verify then with extra efforts. Even with recent prompting strategies that provide rationales before making predictions (Wei et al., 2022), language models can still hallucinate statements, produce flawed reasoning, and output wrong answers. Consequently, novel approaches that enable more reliable reasoning are needed urgently. Some potential directions for this include: (i) using language models to provide evidence, such as theorems, to support the reasoning process; (ii) incorporating a mechanism that makes a judgment when the model is unsure of the answer; and (iii) using a model itself or another module to detect and locate mistakes in a model's reasoning." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 688, + 212, + 700 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 688, + 212, + 700 + ], + "spans": [ + { + "bbox": [ + 67, + 688, + 212, + 700 + ], + "type": "text", + "content": "7.3 Learning from Feedback" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 706, + 292, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 706, + 292, + 774 + ], + "spans": [ + { + "bbox": [ + 67, + 706, + 292, + 774 + ], + "type": "text", + "content": "Another important direction to further improve language models for mathematical reasoning is to let the model learn from feedback. Such a process makes the continual improvement of models' output quality and safety possible. An example is us-" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 71, + 527, + 248 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 527, + 248 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 527, + 248 + ], + "type": "text", + "content": "ing reinforcement learning from human feedback (RLHF) (Ouyang et al., 2022) to align language models with instructions. The idea is to let humans rank the generated outputs of language models and use the learned reward function to finetune the language model with policy gradient (Ouyang et al., 2022; Glaese et al., 2022; Qiu et al., 2022a). In the context of mathematical reasoning, feedback does not necessarily come from humans directly. The outcome of a theorem-proof engine (Jiang et al., 2021; Wu et al., 2021d, 2022c) or the execution result of model-generated scripts can also be used as the reward source (Polu and Sutskever, 2020)." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 259, + 509, + 271 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 259, + 509, + 271 + ], + "spans": [ + { + "bbox": [ + 302, + 259, + 509, + 271 + ], + "type": "text", + "content": "7.4 Multi-modal Mathematical Reasoning" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 278, + 527, + 562 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 278, + 527, + 562 + ], + "spans": [ + { + "bbox": [ + 302, + 278, + 527, + 562 + ], + "type": "text", + "content": "In recent years, there has been growing interest in multi-modal mathematical reasoning, which involves using multiple sources of information, such as text, tables, natural images, and diagrams (Kahou et al., 2018; Kafle et al., 2018; Lu et al., 2021b, 2022b). However, currently available datasets in this domain tend to be small (Zhao et al., 2022), generated from templates (Kahou et al., 2018), or focus on specific topics (Lu et al., 2021a; Chen et al., 2022a). One line of current research involves applying VQA-based frameworks to analyze figures and plots, but this approach can result in significant semantic gaps due to the fact that most VQA models are trained on natural images. One potential direction for future work is to enhance the ability of multi-modal mathematical reasoning systems to tackle more complex and realistic problems. This may involve creating unified models for interpreting and integrating different modalities, as well as developing better evaluation benchmarks to assess the performance of these systems." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 574, + 381, + 587 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 574, + 381, + 587 + ], + "spans": [ + { + "bbox": [ + 302, + 574, + 381, + 587 + ], + "type": "text", + "content": "8 Conclusion" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 597, + 527, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 597, + 527, + 773 + ], + "spans": [ + { + "bbox": [ + 302, + 597, + 527, + 773 + ], + "type": "text", + "content": "In this paper, we present a comprehensive survey of deep learning for mathematical reasoning. We review the various tasks, datasets, and deep learning approaches. We also identify several gaps in the existing datasets and methods. Finally, we outline directions for future research and highlight the potential for further exploration in this field. Our goal with this paper is to provide a comprehensive and useful resource for readers interested in the development of deep learning for mathematical reasoning. To aid in this effort, we have created a reading list that will be continually updated in a GitHub repository at https://github.com/lupantech/dl4math." + } + ] + } + ], + "index": 10 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 792 + ], + "type": "text", + "content": "14613" + } + ] + } + ], + "index": 11 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 8 + }, + { + "para_blocks": [ + { + "bbox": [ + 68, + 71, + 131, + 83 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 71, + 131, + 83 + ], + "spans": [ + { + "bbox": [ + 68, + 71, + 131, + 83 + ], + "type": "text", + "content": "Limitations" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 92, + 291, + 321 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 92, + 291, + 321 + ], + "spans": [ + { + "bbox": [ + 69, + 92, + 291, + 321 + ], + "type": "text", + "content": "One limitation of our survey work is that it is focused on the intersection of mathematical reasoning and deep learning over the past decade, which may not encompass the entire field and its history. Additionally, our evaluation of existing benchmarks and methods is based on a curated set of papers and may not fully represent the state of the art in the field. Furthermore, due to the fast-paced nature of the field, our survey may not reflect the latest developments and advancements which may have come out close to or after the survey was conducted. Despite these limitations, our survey still provides a valuable overview of the current state and key trends in the field of mathematical reasoning and deep learning, and can serve as a valuable resource for researchers and practitioners working in this field." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 68, + 333, + 153, + 346 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 333, + 153, + 346 + ], + "spans": [ + { + "bbox": [ + 68, + 333, + 153, + 346 + ], + "type": "text", + "content": "Broader Impact" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 354, + 291, + 611 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 354, + 291, + 611 + ], + "spans": [ + { + "bbox": [ + 69, + 354, + 291, + 611 + ], + "type": "text", + "content": "Our survey paper on the intersection of mathematical reasoning and deep learning has the potential to significantly impact the field of artificial intelligence. By providing a comprehensive overview of the key tasks, datasets, and methods that have been developed in the past decade, we give researchers and practitioners a clear understanding of the current state-of-the-art and help them make informed decisions about their own research. Additionally, by evaluating existing benchmarks and methods and discussing future research directions, we aim to identify gaps in the current state of the art and guide future research and development efforts towards more advanced and effective mathematical reasoning systems. Overall, our survey has the potential to contribute to the advancement of mathematical reasoning and deep learning, and have a profound impact on machine learning and natural language processing." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 68, + 634, + 127, + 646 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 634, + 127, + 646 + ], + "spans": [ + { + "bbox": [ + 68, + 634, + 127, + 646 + ], + "type": "text", + "content": "References" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 652, + 290, + 772 + ], + "type": "list", + "angle": 0, + "index": 7, + "blocks": [ + { + "bbox": [ + 69, + 652, + 290, + 708 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 652, + 290, + 708 + ], + "spans": [ + { + "bbox": [ + 69, + 652, + 290, + 708 + ], + "type": "text", + "content": "Alexander A. Alemi, François Chollet, Niklas Een, Geoffrey Irving, Christian Szegedy, and Josef Urban. 2016. Deepmath - deep sequence models for premise selection. Advances in neural information processing systems (NeurIPS), 29." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 717, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 717, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 717, + 290, + 772 + ], + "type": "text", + "content": "Reem Alghamdi, Zhenwen Liang, and Xiangliang Zhang. 2022. Armath: a dataset for solving arabic math word problems. In Proceedings of the Thirteenth Language Resources and Evaluation Conference (LREC), pages 351-362." + } + ] + } + ], + "index": 6 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 526, + 772 + ], + "type": "list", + "angle": 0, + "index": 19, + "blocks": [ + { + "bbox": [ + 304, + 72, + 525, + 117 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 525, + 117 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 525, + 117 + ], + "type": "text", + "content": "Chris Alvin, Sumit Gulwani, Rupak Majumdar, and Supratik Mukhopadhyay. 2017. Synthesis of solutions for shaded area geometry problems. In *The Thirtieth International Flairs Conference*." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 304, + 125, + 526, + 215 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 125, + 526, + 215 + ], + "spans": [ + { + "bbox": [ + 304, + 125, + 526, + 215 + ], + "type": "text", + "content": "Aida Amini, Saadia Gabriel, Shanchuan Lin, Rik Koncel-Kedziorski, Yejin Choi, and Hannaneh Hajishirzi. 2019. Mathqa: Towards interpretable math word problem solving with operation-based formalisms. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 2357-2367." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 304, + 223, + 526, + 268 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 223, + 526, + 268 + ], + "spans": [ + { + "bbox": [ + 304, + 223, + 526, + 268 + ], + "type": "text", + "content": "Connor Anderson and Ryan Farrell. 2022. Improving fractal pre-training. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 1300-1309." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 304, + 277, + 525, + 343 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 277, + 525, + 343 + ], + "spans": [ + { + "bbox": [ + 304, + 277, + 525, + 343 + ], + "type": "text", + "content": "Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. 2018. Bottom-up and top-down attention for image captioning and visual question answering. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pages 6077-6086." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 304, + 352, + 525, + 419 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 352, + 525, + 419 + ], + "spans": [ + { + "bbox": [ + 304, + 352, + 525, + 419 + ], + "type": "text", + "content": "Cem Anil, Yuhuai Wu, Anders Johan Andreassen, Aitor Lewkowycz, Vedant Misra, Vinay Venkatesh Ramasesh, Ambrose Slone, Guy Gur-Ari, Ethan Dyer, and Behnam Neyshabur. 2022. Exploring length generalization in large language models. In Advances in Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 428, + 525, + 483 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 428, + 525, + 483 + ], + "spans": [ + { + "bbox": [ + 304, + 428, + 525, + 483 + ], + "type": "text", + "content": "Jacob Austin, Augustus Odena, Maxwell Nye, Maarten Bosma, Henryk Michalewski, David Dohan, Ellen Jiang, Carrie Cai, Michael Terry, Quoc Le, et al. 2021. Program synthesis with large language models. arXiv preprint arXiv:2108.07732." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 491, + 525, + 537 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 491, + 525, + 537 + ], + "spans": [ + { + "bbox": [ + 304, + 491, + 525, + 537 + ], + "type": "text", + "content": "Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 545, + 525, + 601 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 545, + 525, + 601 + ], + "spans": [ + { + "bbox": [ + 304, + 545, + 525, + 601 + ], + "type": "text", + "content": "Kshitij Bansal, Sarah Loos, Markus Rabe, Christian Szegedy, and Stewart Wilcox. 2019. Holist: An environment for machine learning of higher order logic theorem proving. In International Conference on Machine Learning (ICML), pages 454-463. PMLR." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 610, + 525, + 665 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 610, + 525, + 665 + ], + "spans": [ + { + "bbox": [ + 304, + 610, + 525, + 665 + ], + "type": "text", + "content": "Bruno Barras, Samuel Boutin, Cristina Cornes, Judicael Courant, Yann Coscoy, David Delahaye, Daniel de Rauglaudre, Jean-Christophe Filliatre, Eduardo Gimenez, Hugo Herbelin, et al. 1999. The coq proof assistant reference manual. INRIA, version, 6(11)." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 674, + 525, + 730 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 674, + 525, + 730 + ], + "spans": [ + { + "bbox": [ + 304, + 674, + 525, + 730 + ], + "type": "text", + "content": "Taylor Berg-Kirkpatrick and Daniel Spokoyny. 2020. An empirical investigation of contextualized number prediction. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4754-4764." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 739, + 525, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 739, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 739, + 525, + 772 + ], + "type": "text", + "content": "Arindam Bhattacharya. 2017. A survey of question answering for math and science problem. arXiv preprint arXiv:1705.04530." + } + ] + } + ], + "index": 18 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14614" + } + ] + } + ], + "index": 20 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 9 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 289, + 772 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 69, + 72, + 289, + 105 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 72, + 289, + 105 + ], + "spans": [ + { + "bbox": [ + 69, + 72, + 289, + 105 + ], + "type": "text", + "content": "Daniel G Bobrow. 1964. Natural language input for a computer problem solving system. AI Technical Reports." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 114, + 289, + 179 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 114, + 289, + 179 + ], + "spans": [ + { + "bbox": [ + 69, + 114, + 289, + 179 + ], + "type": "text", + "content": "Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. Advances in Neural Information Processing Systems (NeurIPS), 33:1877-1901." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 188, + 289, + 243 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 188, + 289, + 243 + ], + "spans": [ + { + "bbox": [ + 69, + 188, + 289, + 243 + ], + "type": "text", + "content": "Jie Cao and Jing Xiao. 2022. An augmented benchmark dataset for geometric question answering through dual parallel text encoding. In Proceedings of the 29th International Conference on Computational Linguistics (COLING), pages 1511-1520." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 252, + 289, + 306 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 252, + 289, + 306 + ], + "spans": [ + { + "bbox": [ + 69, + 252, + 289, + 306 + ], + "type": "text", + "content": "Yixuan Cao, Feng Hong, Hongwei Li, and Ping Luo. 2021. A bottom-up dag structure extraction model for math word problems. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 39-46." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 315, + 289, + 337 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 315, + 289, + 337 + ], + "spans": [ + { + "bbox": [ + 69, + 315, + 289, + 337 + ], + "type": "text", + "content": "François Charton. 2022. Linear algebra with transformers. Transactions on Machine Learning Research." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 345, + 289, + 412 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 345, + 289, + 412 + ], + "spans": [ + { + "bbox": [ + 69, + 345, + 289, + 412 + ], + "type": "text", + "content": "Jiaqi Chen, Tong Li, Jinghui Qin, Pan Lu, Liang Lin, Chongyu Chen, and Xiaodan Liang. 2022a. Unigeo: Unifying geometry logical reasoning via reformulating mathematical expression. In The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP)." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 420, + 289, + 486 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 420, + 289, + 486 + ], + "spans": [ + { + "bbox": [ + 69, + 420, + 289, + 486 + ], + "type": "text", + "content": "Jiaqi Chen, Jianheng Tang, Jinghui Qin, Xiaodan Liang, Lingbo Liu, Eric Xing, and Liang Lin. 2021a. Geoqa: A geometric question answering benchmark towards multimodal numerical reasoning. In Findings of the Association for Computational Linguistics (ACL), pages 513-523." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 494, + 289, + 560 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 494, + 289, + 560 + ], + "spans": [ + { + "bbox": [ + 69, + 494, + 289, + 560 + ], + "type": "text", + "content": "Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, et al. 2021b. Evaluating large language models trained on code. arXiv preprint arXiv:2107.03374." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 568, + 289, + 623 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 568, + 289, + 623 + ], + "spans": [ + { + "bbox": [ + 69, + 568, + 289, + 623 + ], + "type": "text", + "content": "Wenhu Chen, Xueguang Ma, Xinyi Wang, and William W Cohen. 2022b. Program of thoughts prompting: Disentangling computation from reasoning for numerical reasoning tasks. arXiv preprint arXiv:2211.12588." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 631, + 289, + 686 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 631, + 289, + 686 + ], + "spans": [ + { + "bbox": [ + 69, + 631, + 289, + 686 + ], + "type": "text", + "content": "Wenhu Chen, Ming Yin, Max Ku, Elaine Wan, Xueguang Ma, Jianyu Xu, Tony Xia, Xinyi Wang, and Pan Lu. 2023. Theoremq: A theorem-driven question answering dataset. arXiv preprint arXiv:2305.12524." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 694, + 289, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 694, + 289, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 694, + 289, + 772 + ], + "type": "text", + "content": "Zhiyu Chen, Wenhu Chen, Charese Smiley, Sameena Shah, Iana Borova, Dylan Langdon, Reema Moussa, Matt Beane, Ting-Hao Huang, Bryan R Routledge, et al. 2021c. Finqa: A dataset of numerical reasoning over financial data. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3697-3711." + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 524, + 772 + ], + "type": "list", + "angle": 0, + "index": 22, + "blocks": [ + { + "bbox": [ + 304, + 72, + 524, + 127 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 524, + 127 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 524, + 127 + ], + "type": "text", + "content": "Zhiyu Chen, Shiyang Li, Charese Smiley, Zhiqiang Ma, Sameena Shah, and William Yang Wang. 2022c. Convfinqa: Exploring the chain of numerical reasoning in conversational finance question answering. arXiv preprint arXiv:2210.03849." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 136, + 524, + 212 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 136, + 524, + 212 + ], + "spans": [ + { + "bbox": [ + 304, + 136, + 524, + 212 + ], + "type": "text", + "content": "Ting-Rui Chiang and Yun-Nung Chen. 2019. Semantically-aligned equation generation for solving and reasoning math word problems. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 2656-2668." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 221, + 524, + 275 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 221, + 524, + 275 + ], + "spans": [ + { + "bbox": [ + 304, + 221, + 524, + 275 + ], + "type": "text", + "content": "Jaemin Cho, Jie Lei, Hao Tan, and Mohit Bansal. 2021. Unifying vision-and-language tasks via text generation. In Proceedings of the 38th International Conference on Machine Learning (ICML), pages 1931-1942." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 283, + 524, + 361 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 283, + 524, + 361 + ], + "spans": [ + { + "bbox": [ + 304, + 283, + 524, + 361 + ], + "type": "text", + "content": "Kyunghyun Cho, Bart van Merrienboer Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using rn encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1724-1734." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 368, + 524, + 413 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 368, + 524, + 413 + ], + "spans": [ + { + "bbox": [ + 304, + 368, + 524, + 413 + ], + "type": "text", + "content": "Shang-Ching Chou, Xiao-Shan Gao, and Jing-Zhong Zhang. 1996. Automated generation of readable proofs with geometric invariants. Journal of Automated Reasoning, 17(3):325-347." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 421, + 524, + 486 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 421, + 524, + 486 + ], + "spans": [ + { + "bbox": [ + 304, + 421, + 524, + 486 + ], + "type": "text", + "content": "Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, et al. 2022. Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 495, + 524, + 560 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 495, + 524, + 560 + ], + "spans": [ + { + "bbox": [ + 304, + 495, + 524, + 560 + ], + "type": "text", + "content": "Peter Clark, Oren Etzioni, Tushar Khot, Daniel Khashabi, Bhavana Mishra, Kyle Richardson, Ashish Sabharwal, Carissa Schoenick, Oyvind Tafjord, Niket Tandon, et al. 2020. From 'f'to 'a'on the ny regents science exams: An overview of the aristo project. AI Magazine, 41(4):39-53." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 568, + 524, + 623 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 568, + 524, + 623 + ], + "spans": [ + { + "bbox": [ + 304, + 568, + 524, + 623 + ], + "type": "text", + "content": "Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. 2021. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 632, + 524, + 709 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 632, + 524, + 709 + ], + "spans": [ + { + "bbox": [ + 304, + 632, + 524, + 709 + ], + "type": "text", + "content": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 4171-4186." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 717, + 524, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 717, + 524, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 717, + 524, + 772 + ], + "type": "text", + "content": "Dheeru Dua, Yizhong Wang, Pradeep Dasigi, Gabriel Stanovsky, Sameer Singh, and Matt Gardner. 2019. Drop: A reading comprehension benchmark requiring discrete reasoning over paragraphs. In Proceedings of the 2019 Conference of the North American" + } + ] + } + ], + "index": 21 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14615" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 10 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 772 + ], + "type": "list", + "angle": 0, + "index": 12, + "blocks": [ + { + "bbox": [ + 80, + 72, + 291, + 105 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 72, + 291, + 105 + ], + "spans": [ + { + "bbox": [ + 80, + 72, + 291, + 105 + ], + "type": "text", + "content": "Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 2368-2378." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 117, + 289, + 139 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 117, + 289, + 139 + ], + "spans": [ + { + "bbox": [ + 69, + 117, + 289, + 139 + ], + "type": "text", + "content": "Edward A Feigenbaum et al. 1963. Computers and thought. McGraw-Hill." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 70, + 150, + 290, + 195 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 70, + 150, + 290, + 195 + ], + "spans": [ + { + "bbox": [ + 70, + 150, + 290, + 195 + ], + "type": "text", + "content": "Yu Feng, Jing Zhang, Xiaokang Zhang, Lemao Liu, Cuiping Li, and Hong Chen. 2021. Injecting numerical reasoning skills into knowledge base question answering models. arXiv preprint arXiv:2112.06109." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 206, + 290, + 261 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 206, + 290, + 261 + ], + "spans": [ + { + "bbox": [ + 69, + 206, + 290, + 261 + ], + "type": "text", + "content": "Deborah Ferreira and André Freitas. 2020a. Natural language premise selection: Finding supporting statements for mathematical text. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pages 2175-2182." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 272, + 290, + 327 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 272, + 290, + 327 + ], + "spans": [ + { + "bbox": [ + 69, + 272, + 290, + 327 + ], + "type": "text", + "content": "Deborah Ferreira and André Freitas. 2020b. Premise selection in natural language mathematical texts. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 7365-7374." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 338, + 289, + 384 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 338, + 289, + 384 + ], + "spans": [ + { + "bbox": [ + 69, + 338, + 289, + 384 + ], + "type": "text", + "content": "Yao Fu, Hao Peng, Ashish Sabharwal, Peter Clark, and Tushar Khot. 2023. Complexity-based prompting for multi-step reasoning. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 395, + 290, + 450 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 395, + 290, + 450 + ], + "spans": [ + { + "bbox": [ + 69, + 395, + 290, + 450 + ], + "type": "text", + "content": "Leo Gao, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, Horace He, Anish Thite, Noa Nabeshima, et al. 2020. The pile: An 800gb dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 461, + 290, + 506 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 461, + 290, + 506 + ], + "spans": [ + { + "bbox": [ + 69, + 461, + 290, + 506 + ], + "type": "text", + "content": "Luyu Gao, Aman Madaan, Shuyan Zhou, Uri Alon, Pengfei Liu, Yiming Yang, Jamie Callan, and Graham Neubig. 2022. Pal: Program-aided language models. arXiv preprint arXiv:2211.10435." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 517, + 290, + 583 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 517, + 290, + 583 + ], + "spans": [ + { + "bbox": [ + 69, + 517, + 290, + 583 + ], + "type": "text", + "content": "Peng Gao, Zhengkai Jiang, Haoxuan You, Pan Lu, Steven CH Hoi, Xiaogang Wang, and Hongsheng Li. 2019. Dynamic fusion with intra-and inter-modality attention flow for visual question answering. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 6639-6648." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 594, + 290, + 640 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 594, + 290, + 640 + ], + "spans": [ + { + "bbox": [ + 69, + 594, + 290, + 640 + ], + "type": "text", + "content": "Thibault Gauthier, Cezary Kaliszyk, Josef Urban, Ramana Kumar, and Michael Norrish. 2021. TacticToe: Learning to Prove with Tactics. Journal of Automated Reasoning." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 650, + 290, + 704 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 650, + 290, + 704 + ], + "spans": [ + { + "bbox": [ + 69, + 650, + 290, + 704 + ], + "type": "text", + "content": "Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, and Yann N Dauphin. 2017. Convolutional sequence to sequence learning. In International conference on machine learning (ICML), pages 1243-1252. PMLR." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 69, + 716, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 716, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 716, + 290, + 772 + ], + "type": "text", + "content": "Herbert Gelernter, James R Hansen, and Donald W Loveland. 1960. Empirical explorations of the geometry theorem machine. In Papers presented at the May 3-5, 1960, western joint IRE-AIEE-ACM computer conference, pages 143-149." + } + ] + } + ], + "index": 11 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 525, + 772 + ], + "type": "list", + "angle": 0, + "index": 24, + "blocks": [ + { + "bbox": [ + 304, + 72, + 525, + 127 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 525, + 127 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 525, + 127 + ], + "type": "text", + "content": "Mor Geva, Ankit Gupta, and Jonathan Berant. 2020. Injecting numerical reasoning skills into language models. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 946-958." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 137, + 525, + 192 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 137, + 525, + 192 + ], + "spans": [ + { + "bbox": [ + 304, + 137, + 525, + 192 + ], + "type": "text", + "content": "Kevin Gimpel, Dipanjan Das, and Noah A Smith. 2010. Distributed asynchronous online learning for natural language processing. In Proceedings of the Fourteenth Conference on Computational Natural Language Learning, pages 213-222." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 201, + 525, + 266 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 201, + 525, + 266 + ], + "spans": [ + { + "bbox": [ + 304, + 201, + 525, + 266 + ], + "type": "text", + "content": "Amelia Glaese, Nat McAleese, Maja Trèbacz, John Aslanides, Vlad Firoiu, Timo Ewalds, Maribeth Rauh, Laura Weidinger, Martin Chadwick, Phoebe Thacker, et al. 2022. Improving alignment of dialogue agents via targeted human judgements. arXiv preprint arXiv:2209.14375." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 277, + 525, + 310 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 277, + 525, + 310 + ], + "spans": [ + { + "bbox": [ + 304, + 277, + 525, + 310 + ], + "type": "text", + "content": "Adam Grabowski, Artur Korniłowicz, and Adam Naumowicz. 2015. Four decades of mizar. Journal of Automated Reasoning, 55(3):191-198." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 319, + 525, + 373 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 319, + 525, + 373 + ], + "spans": [ + { + "bbox": [ + 304, + 319, + 525, + 373 + ], + "type": "text", + "content": "Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Mingwei Chang. 2020. Retrieval augmented language model pre-training. In International Conference on Machine Learning (ICML), pages 3929-3938. PMLR." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 384, + 525, + 439 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 384, + 525, + 439 + ], + "spans": [ + { + "bbox": [ + 304, + 384, + 525, + 439 + ], + "type": "text", + "content": "Jesse Michael Han, Jason Rute, Yuhuai Wu, Edward W Ayers, and Stanislas Polu. 2022. Proof artifact constraining for theorem proving with language models. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 448, + 525, + 492 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 448, + 525, + 492 + ], + "spans": [ + { + "bbox": [ + 304, + 448, + 525, + 492 + ], + "type": "text", + "content": "Yihan Hao, Mingliang Zhang, Fei Yin, and Linlin Huang. 2022. Pgdp5k: A diagram parsing dataset for plane geometry problems. In 26th International Conference on Pattern Recognition (ICPR)." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 502, + 525, + 555 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 502, + 525, + 555 + ], + "spans": [ + { + "bbox": [ + 304, + 502, + 525, + 555 + ], + "type": "text", + "content": "Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pages 770-778." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 565, + 525, + 621 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 565, + 525, + 621 + ], + "spans": [ + { + "bbox": [ + 304, + 565, + 525, + 621 + ], + "type": "text", + "content": "Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, and Jacob Steinhardt. 2021a. Measuring massive multitask language understanding. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 304, + 630, + 525, + 696 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 630, + 525, + 696 + ], + "spans": [ + { + "bbox": [ + 304, + 630, + 525, + 696 + ], + "type": "text", + "content": "Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, and Jacob Steinhardt. 2021b. Measuring mathematical problem solving with the math dataset. In 35th Conference on Neural Information Processing Systems (NeurIPS) Track on Datasets and Benchmarks." + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 304, + 706, + 525, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 706, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 706, + 525, + 772 + ], + "type": "text", + "content": "Dan Hendrycks, Xiaoyuan Liu, Eric Wallace, Adam Dziedzic, Rishabh Krishnan, and Dawn Song. 2020. Pretrained transformers improve out-of-distribution robustness. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 2744-2751." + } + ] + } + ], + "index": 23 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14616" + } + ] + } + ], + "index": 25 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 11 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 772 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "spans": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "type": "text", + "content": "Jonathan Herzig, Pawel Krzysztof Nowak, Thomas Mueller, Francesco Piccinno, and Julian Eisenschlos. 2020. Tapas: Weakly supervised table parsing via pre-training. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 4320-4333." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 148, + 290, + 181 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 148, + 290, + 181 + ], + "spans": [ + { + "bbox": [ + 69, + 148, + 290, + 181 + ], + "type": "text", + "content": "Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation, 9(8):1735-1780." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 192, + 290, + 248 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 192, + 290, + 248 + ], + "spans": [ + { + "bbox": [ + 69, + 192, + 290, + 248 + ], + "type": "text", + "content": "Yining Hong, Qing Li, Daniel Ciao, Siyuan Huang, and Song-Chun Zhu. 2021a. Learning by fixing: Solving math word problems with weak supervision. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 4959-4967." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 258, + 290, + 302 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 258, + 290, + 302 + ], + "spans": [ + { + "bbox": [ + 69, + 258, + 290, + 302 + ], + "type": "text", + "content": "Yining Hong, Qing Li, Ran Gong, Daniel Ciao, Siyuan Huang, and Song-Chun Zhu. 2021b. Smart: A situation model for algebra story problems via attributed grammar. In AAAI, pages 13009-13017." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 312, + 290, + 368 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 312, + 290, + 368 + ], + "spans": [ + { + "bbox": [ + 69, + 312, + 290, + 368 + ], + "type": "text", + "content": "Mohammad Javad Hosseini, Hannaneh Hajishirzi, Oren Etzioni, and Nate Kushman. 2014. Learning to solve arithmetic word problems with verb categorization. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 378, + 290, + 423 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 378, + 290, + 423 + ], + "spans": [ + { + "bbox": [ + 69, + 378, + 290, + 423 + ], + "type": "text", + "content": "Daniel Huang, Prafulla Dhariwal, Dawn Song, and Ilya Sutskever. 2019. Gamepad: A learning environment for theorem proving. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 433, + 290, + 488 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 433, + 290, + 488 + ], + "spans": [ + { + "bbox": [ + 69, + 433, + 290, + 488 + ], + "type": "text", + "content": "Danqing Huang, Jing Liu, Chin-Yew Lin, and Jian Yin. 2018. Neural math word problem solver with reinforcement learning. In Proceedings of the 27th International Conference on Computational Linguistics (COLING), pages 213-223." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 498, + 290, + 554 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 498, + 290, + 554 + ], + "spans": [ + { + "bbox": [ + 69, + 498, + 290, + 554 + ], + "type": "text", + "content": "Danqing Huang, Shuming Shi, Chin-Yew Lin, and Jian Yin. 2017. Learning fine-grained expressions to solve math word problems. In Proceedings of Empirical Methods in Natural Language Processing (EMNLP), pages 805-814." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 564, + 290, + 630 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 564, + 290, + 630 + ], + "spans": [ + { + "bbox": [ + 69, + 564, + 290, + 630 + ], + "type": "text", + "content": "Danqing Huang, Shuming Shi, Chin-Yew Lin, Jian Yin, and Wei-Ying Ma. 2016. How well do computers solve math word problems? large-scale dataset construction and evaluation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL), pages 887-896." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 640, + 290, + 717 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 640, + 290, + 717 + ], + "spans": [ + { + "bbox": [ + 69, + 640, + 290, + 717 + ], + "type": "text", + "content": "Albert Q. Jiang, Sean Welleck, Jin Peng Zhou, Wenda Li, Jiacheng Liu, Mateja Jamnik, Timothée Lacroix, Yuhuai Wu, and Guillaume Lample. 2022a. Draft, sketch, and prove: Guiding formal theorem provers with informal proofs. In Submitted to The Eleventh International Conference on Learning Representations." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 728, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 728, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 728, + 290, + 772 + ], + "type": "text", + "content": "Albert Qiaochu Jiang, Wenda Li, Jesse Michael Han, and Yuhuai Wu. 2021. Lisa: Language models of isabelle proofs. In 6th Conference on Artificial Intelligence and Theorem Proving (AITP)." + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 525, + 772 + ], + "type": "list", + "angle": 0, + "index": 22, + "blocks": [ + { + "bbox": [ + 304, + 72, + 525, + 148 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 525, + 148 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 525, + 148 + ], + "type": "text", + "content": "Albert Qiaochu Jiang, Wenda Li, Szymon Tworkowski, Konrad Czechowski, Tomasz Odrzygoźdź, Piotr Miłos, Yuhuai Wu, and Mateja Jamnik. 2022b. Thor: Wielding hammers to integrate language models and automated theorem provers. Advances in Neural Information Processing Systems (NeurIPS), 35:8360-8373." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 158, + 525, + 214 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 158, + 525, + 214 + ], + "spans": [ + { + "bbox": [ + 304, + 158, + 525, + 214 + ], + "type": "text", + "content": "Zhanming Jie, Jierui Li, and Wei Lu. 2022. Learning to reason deductively: Math word problem solving as complex relation extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL), pages 5944-5955." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 223, + 525, + 300 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 223, + 525, + 300 + ], + "spans": [ + { + "bbox": [ + 304, + 223, + 525, + 300 + ], + "type": "text", + "content": "Jaehun Jung, Lianhui Qin, Sean Welleck, Faeze Brahman, Chandra Bhagavatula, Ronan Le Bras, and Yejin Choi. 2022. Maieutic prompting: Logically consistent reasoning with recursive explanations. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1266-1279." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 309, + 525, + 364 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 309, + 525, + 364 + ], + "spans": [ + { + "bbox": [ + 304, + 309, + 525, + 364 + ], + "type": "text", + "content": "Kushal Kafle, Brian Price, Scott Cohen, and Christopher Kanan. 2018. Dvqa: Understanding data visualizations via question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5648-5656." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 374, + 525, + 428 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 374, + 525, + 428 + ], + "spans": [ + { + "bbox": [ + 304, + 374, + 525, + 428 + ], + "type": "text", + "content": "Samira Ebrahimi Kahou, Vincent Michalski, Adam Atkinson, Ákos Kárdár, Adam Trischler, and Yoshua Bengio. 2018. Figureqa: An annotated figure dataset for visual reasoning. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 438, + 525, + 492 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 438, + 525, + 492 + ], + "spans": [ + { + "bbox": [ + 304, + 438, + 525, + 492 + ], + "type": "text", + "content": "Cezary Kaliszyk, François Chollet, and Christian Szegedy. 2017. Holstep: A machine learning dataset for higher-order logic theorem proving. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 502, + 525, + 579 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 502, + 525, + 579 + ], + "spans": [ + { + "bbox": [ + 304, + 502, + 525, + 579 + ], + "type": "text", + "content": "Ashwin Kalyan, Abhinav Kumar, Arjun Chandrasekaran, Ashish Sabharwal, and Peter Clark. 2021. How much coffee was consumed during emnlp 2019? fermi problems: A new reasoning challenge for ai. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7318-7328." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 588, + 525, + 632 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 588, + 525, + 632 + ], + "spans": [ + { + "bbox": [ + 304, + 588, + 525, + 632 + ], + "type": "text", + "content": "Nikhil Kandpal, H. Deng, Adam Roberts, Eric Wallace, and Colin Raffel. 2022. Large language models struggle to learn long-tail knowledge. *ArXiv*, abs/2211.08411." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 641, + 525, + 708 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 641, + 525, + 708 + ], + "spans": [ + { + "bbox": [ + 304, + 641, + 525, + 708 + ], + "type": "text", + "content": "Daniel Khashabi, Sewon Min, Tushar Khot, Ashish Sabharwal, Oyvind Tafjord, Peter Clark, and Hannaneh Hajishirzi. 2020. Unifiedqa: Crossing format boundaries with a single qa system. In Findings of the Association for Computational Linguistics (EMNLP), pages 1896-1907." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 717, + 525, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 717, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 717, + 525, + 772 + ], + "type": "text", + "content": "Tushar Khot, Harsh Trivedi, Matthew Finlayson, Yao Fu, Kyle Richardson, Peter Clark, and Ashish Sabharwal. 2022. Decomposed prompting: A modular approach for solving complex tasks. arXiv preprint arXiv:2210.02406." + } + ] + } + ], + "index": 21 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14617" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 12 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 772 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "spans": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "type": "text", + "content": "Bugeun Kim, Kyung Seo Ki, Donggeon Lee, and Gahgene Gweon. 2020. Point to the expression: Solving algebraic word problems using the expression-pointer transformer model. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3768-3779." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 145, + 290, + 190 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 145, + 290, + 190 + ], + "spans": [ + { + "bbox": [ + 69, + 145, + 290, + 190 + ], + "type": "text", + "content": "Jin-Hwa Kim, Jaehyun Jun, and Byoung-Tak Zhang. 2018. Bilinear attention networks. In Advances in Neural Information Processing Systems (NeurIPS), pages 1571-1581." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 196, + 290, + 253 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 196, + 290, + 253 + ], + "spans": [ + { + "bbox": [ + 69, + 196, + 290, + 253 + ], + "type": "text", + "content": "Wonjae Kim, Bokyung Son, and Ildoo Kim. 2021. Vilt: Vision-and-language transformer without convolution or region supervision. In Proceedings of the 38th International Conference on Machine Learning (ICML), pages 5583-5594." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 259, + 290, + 314 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 259, + 290, + 314 + ], + "spans": [ + { + "bbox": [ + 69, + 259, + 290, + 314 + ], + "type": "text", + "content": "Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. In 36th Conference on Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 321, + 290, + 397 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 321, + 290, + 397 + ], + "spans": [ + { + "bbox": [ + 69, + 321, + 290, + 397 + ], + "type": "text", + "content": "Rik Koncel-K., Subhro Roy, Aida Amini, Nate Kushman, and Hannaneh Hajishirzi. 2016. Mawps: A math word problem repository. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL), pages 1152-1157." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 405, + 290, + 460 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 405, + 290, + 460 + ], + "spans": [ + { + "bbox": [ + 69, + 405, + 290, + 460 + ], + "type": "text", + "content": "Rik Koncel-Kedziorski, Hannaneh Hajishirzi, Ashish Sabharwal, Oren Etzioni, and Siena Dumas Ang. 2015. Parsing algebraic word problems into equations. Transactions of the Association for Computational Linguistics (TACL), 3:585-597." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 467, + 290, + 522 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 467, + 290, + 522 + ], + "spans": [ + { + "bbox": [ + 69, + 467, + 290, + 522 + ], + "type": "text", + "content": "Kundan Krishna, Jeffrey Bigham, and Zachary C Lipton. 2021. Does pretraining for summarization require knowledge transfer? In *Findings of the Association for Computational Linguistics: EMNLP* 2021, pages 3178-3189." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 530, + 290, + 585 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 530, + 290, + 585 + ], + "spans": [ + { + "bbox": [ + 69, + 530, + 290, + 585 + ], + "type": "text", + "content": "Nate Kushman, Yoav Artzi, Luke Zettlemoyer, and Regina Barzilay. 2014. Learning to automatically solve algebra word problems. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (ACL), pages 271-281." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 592, + 290, + 626 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 592, + 290, + 626 + ], + "spans": [ + { + "bbox": [ + 69, + 592, + 290, + 626 + ], + "type": "text", + "content": "Guillaume Lample and François Charton. 2020. Deep learning for symbolic mathematics. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 633, + 290, + 698 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 633, + 290, + 698 + ], + "spans": [ + { + "bbox": [ + 69, + 633, + 290, + 698 + ], + "type": "text", + "content": "Guillaume Lample, Timothee Lacroix, Marie-Anne Lachaux, Aurelien Rodriguez, Amaury Hayat, Thibaut Lavril, Gabriel Ebner, and Xavier Martinet. 2022. Hypertree proof search for neural theorem proving. Advances in Neural Information Processing Systems (NeurIPS), 35:26337-26349." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 706, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 706, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 706, + 290, + 772 + ], + "type": "text", + "content": "Yihuai Lan, Lei Wang, Qiyuan Zhang, Yunshi Lan, Bing Tian Dai, Yan Wang, Dongxiang Zhang, and Ee-Peng Lim. 2022. Mwptoolkit: an open-source framework for deep learning-based math word problem solvers. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 13188-13190." + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 525, + 772 + ], + "type": "list", + "angle": 0, + "index": 22, + "blocks": [ + { + "bbox": [ + 304, + 72, + 525, + 126 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 525, + 126 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 525, + 126 + ], + "type": "text", + "content": "Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 138, + 525, + 181 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 138, + 525, + 181 + ], + "spans": [ + { + "bbox": [ + 304, + 138, + 525, + 181 + ], + "type": "text", + "content": "Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278-2324." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 192, + 525, + 280 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 192, + 525, + 280 + ], + "spans": [ + { + "bbox": [ + 304, + 192, + 525, + 280 + ], + "type": "text", + "content": "Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 7871-7880." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 291, + 525, + 368 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 291, + 525, + 368 + ], + "spans": [ + { + "bbox": [ + 304, + 291, + 525, + 368 + ], + "type": "text", + "content": "Aitor Lewkowycz, Anders Johan Andreassen, David Dohan, Ethan Dyer, Henryk Michalewski, Vinay Venkatesh Ramasesh, Ambrose Slone, Cem Anil, Imanol Schlag, Theo Gutman-Solo, et al. 2022. Solving quantitative reasoning problems with language models. In Advances in Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 378, + 525, + 445 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 378, + 525, + 445 + ], + "spans": [ + { + "bbox": [ + 304, + 378, + 525, + 445 + ], + "type": "text", + "content": "Jierui Li, Lei Wang, Jipeng Zhang, Yan Wang, Bing Tian Dai, and Dongxiang Zhang. 2019. Modeling intra-relation in math word problems with different functional multi-head attentions. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), pages 6162-6167." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 454, + 525, + 510 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 454, + 525, + 510 + ], + "spans": [ + { + "bbox": [ + 304, + 454, + 525, + 510 + ], + "type": "text", + "content": "Jiwei Li, Alexander H Miller, Sumit Chopra, Marc'Aurelio Ranzato, and Jason Weston. 2017. Dialogue learning with human-in-the-loop. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 520, + 525, + 576 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 520, + 525, + 576 + ], + "spans": [ + { + "bbox": [ + 304, + 520, + 525, + 576 + ], + "type": "text", + "content": "Lianian Harold Li, Mark Yatskar, Da Yin, Cho-Jui Hsieh, and Kai-Wei Chang. 2020a. What does bert with vision look at? In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 5265-5275." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 586, + 525, + 663 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 586, + 525, + 663 + ], + "spans": [ + { + "bbox": [ + 304, + 586, + 525, + 663 + ], + "type": "text", + "content": "Shucheng Li, Lingfei Wu, Shiwei Feng, Fangli Xu, Fengyuan Xu, and Sheng Zhong. 2020b. Graph-to-tree neural networks for learning structured input-output translation with applications to semantic parsing and math word problem. In *Findings of the Association for Computational Linguistics (EMNLP)* pages 2841-2852." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 673, + 525, + 718 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 673, + 525, + 718 + ], + "spans": [ + { + "bbox": [ + 304, + 673, + 525, + 718 + ], + "type": "text", + "content": "Wenda Li, Lei Yu, Yuhuai Wu, and Lawrence C Paulson. 2021. Isarstep: a benchmark for high-level mathematical reasoning. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 728, + 525, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 728, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 728, + 525, + 772 + ], + "type": "text", + "content": "Yifei Li, Zeqi Lin, Shizhuo Zhang, Qiang Fu, Bei Chen, Jian-Guang Lou, and Weizhu Chen. 2022a. On the advance of making language models better reasoners. arXiv preprint arXiv:2206.02336." + } + ] + } + ], + "index": 21 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14618" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 13 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 772 + ], + "type": "list", + "angle": 0, + "index": 10, + "blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "spans": [ + { + "bbox": [ + 69, + 72, + 291, + 139 + ], + "type": "text", + "content": "Zhongli Li, Wenxuan Zhang, Chao Yan, Qingyu Zhou, Chao Li, Hongzhi Liu, and Yunbo Cao. 2022b. Seeking patterns, not just memorizing procedures: Contrastive learning for solving math word problems. In *Findings of the Association for Computational Linguistics (ACL)*, pages 2486-2496." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 146, + 290, + 201 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 146, + 290, + 201 + ], + "spans": [ + { + "bbox": [ + 69, + 146, + 290, + 201 + ], + "type": "text", + "content": "Percy Liang, Rishi Bommasani, Tony Lee, Dimitris Tsipras, Dilara Soylu, Michihiro Yasunaga, Yian Zhang, Deepak Narayanan, Yuhuai Wu, Ananya Kumar, et al. 2022a. Holistic evaluation of language models. arXiv preprint arXiv:2211.09110." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 210, + 290, + 265 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 210, + 290, + 265 + ], + "spans": [ + { + "bbox": [ + 69, + 210, + 290, + 265 + ], + "type": "text", + "content": "Percy Liang and Dan Klein. 2009. Online em for unsupervised models. In Proceedings of human language technologies: The 2009 annual conference of the North American chapter of the association for computational linguistics (NAACL), pages 611-619." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 272, + 290, + 338 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 272, + 290, + 338 + ], + "spans": [ + { + "bbox": [ + 69, + 272, + 290, + 338 + ], + "type": "text", + "content": "Zhenwen Liang, Jipeng Zhang, Lei Wang, Wei Qin, Yunshi Lan, Jie Shao, and Xiangliang Zhang. 2022b. Mwp-bert: Numeracy-augmented pre-training for math word problem solving. In *Findings of the Association for Computational Linguistics (NAACL)* pages 997-1009." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 346, + 290, + 413 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 346, + 290, + 413 + ], + "spans": [ + { + "bbox": [ + 69, + 346, + 290, + 413 + ], + "type": "text", + "content": "Bill Yuchen Lin, Seyeon Lee, Rahul Khanna, and Xi-ang Ren. 2020. Birds have four legs?! numersense: Probing numerical commonsense knowledge of pretrained language models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6862-6868." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 420, + 290, + 487 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 420, + 290, + 487 + ], + "spans": [ + { + "bbox": [ + 69, + 420, + 290, + 487 + ], + "type": "text", + "content": "Xin Lin, Zhenya Huang, Hongke Zhao, Enhong Chen, Qi Liu, Hao Wang, and Shijin Wang. 2021. Hms: A hierarchical solver with dependency-enhanced understanding for math word problem. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 4232-4240." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 494, + 290, + 561 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 494, + 290, + 561 + ], + "spans": [ + { + "bbox": [ + 69, + 494, + 290, + 561 + ], + "type": "text", + "content": "Wang Ling, Dani Yogatama, Chris Dyer, and Phil Blun-som. 2017. Program induction by rationale generation: Learning to solve and explain algebraic word problems. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), pages 158-167." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 568, + 290, + 646 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 568, + 290, + 646 + ], + "spans": [ + { + "bbox": [ + 69, + 568, + 290, + 646 + ], + "type": "text", + "content": "Jiachang Liu, Dinghan Shen, Yizhe Zhang, William B Dolan, Lawrence Carin, and Weizhu Chen. 2022a. What makes good in-context examples for gpt-3? In Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 100-114." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 654, + 290, + 709 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 654, + 290, + 709 + ], + "spans": [ + { + "bbox": [ + 69, + 654, + 290, + 709 + ], + "type": "text", + "content": "Qian Liu, Bei Chen, Jiaqi Guo, Morteza Ziyadi, Zeqi Lin, Weizhu Chen, and Jian-Guang Lou. 2022b. TAPEX: Table pre-training via learning a neural SQL executor. In International Conference on Learning Representations." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 716, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 716, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 716, + 290, + 772 + ], + "type": "text", + "content": "Qianying Liu, Wenyu Guan, Sujian Li, Fei Cheng, Daisuke Kawahara, and Sadao Kurohashi. 2020. Reverse operation based data augmentation for solving math word problems. IEEE Transactions on Audio, Speech and Language Processing." + } + ] + } + ], + "index": 9 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 524, + 772 + ], + "type": "list", + "angle": 0, + "index": 21, + "blocks": [ + { + "bbox": [ + 304, + 72, + 524, + 149 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 524, + 149 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 524, + 149 + ], + "type": "text", + "content": "Qianying Liu, Wenyv Guan, Sujian Li, and Daisuke Kawahara. 2019a. Tree-structured decoding for solving math word problems. In Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pages 2370-2379." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 304, + 158, + 524, + 246 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 158, + 524, + 246 + ], + "spans": [ + { + "bbox": [ + 304, + 158, + 524, + 246 + ], + "type": "text", + "content": "Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019b. Roberta: A robustly optimized bert pretraining approach. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT)." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 255, + 524, + 289 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 255, + 524, + 289 + ], + "spans": [ + { + "bbox": [ + 304, + 255, + 524, + 289 + ], + "type": "text", + "content": "Sarah Loos, Geoffrey Irving, Christian Szegedy, and Cezary Kaliszyk. 2017. Deep network guided proof search. arXiv preprint arXiv:1701.06972." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 297, + 524, + 364 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 297, + 524, + 364 + ], + "spans": [ + { + "bbox": [ + 304, + 297, + 524, + 364 + ], + "type": "text", + "content": "Pan Lu, Ran Gong, Shibiao Jiang, Liang Qiu, Siyuan Huang, Xiaodan Liang, and Song-Chun Zhu. 2021a. Inter-gps: Interpretable geometry problem solving with formal language and symbolic reasoning. In The 59th Annual Meeting of the Association for Computational Linguistics (ACL)." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 373, + 524, + 439 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 373, + 524, + 439 + ], + "spans": [ + { + "bbox": [ + 304, + 373, + 524, + 439 + ], + "type": "text", + "content": "Pan Lu, Swaroop Mishra, Tony Xia, Liang Qiu, Kai-Wei Chang, Song-Chun Zhu, Oyvind Tafjord, Peter Clark, and Ashwin Kalyan. 2022a. Learn to explain: Multimodal reasoning via thought chains for science question answering. In The 36th Conference on Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 448, + 524, + 504 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 448, + 524, + 504 + ], + "spans": [ + { + "bbox": [ + 304, + 448, + 524, + 504 + ], + "type": "text", + "content": "Pan Lu, Baolin Peng, Hao Cheng, Michel Galley, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, and Jianfeng Gao. 2023. Chameleon: Plug-and-play compositional reasoning with large language models. arXiv preprint arXiv:2304.09842." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 513, + 524, + 579 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 513, + 524, + 579 + ], + "spans": [ + { + "bbox": [ + 304, + 513, + 524, + 579 + ], + "type": "text", + "content": "Pan Lu, Liang Qiu, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, Tanmay Rajpurohit, Peter Clark, and Ashwin Kalyan. 2022b. Dynamic prompt learning via policy gradient for semi-structured mathematical reasoning. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 587, + 524, + 665 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 587, + 524, + 665 + ], + "spans": [ + { + "bbox": [ + 304, + 587, + 524, + 665 + ], + "type": "text", + "content": "Pan Lu, Liang Qiu, Jiaqi Chen, Tony Xia, Yizhou Zhao, Wei Zhang, Zhou Yu, Xiaodan Liang, and Song-Chun Zhu. 2021b. Iconqa: A new benchmark for abstract diagram understanding and visual language reasoning. In The 35th Conference on Neural Information Processing Systems (NeurIPS) Track on Datasets and Benchmarks." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 674, + 524, + 740 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 674, + 524, + 740 + ], + "spans": [ + { + "bbox": [ + 304, + 674, + 524, + 740 + ], + "type": "text", + "content": "Yao Lu, Max Bartolo, Alastair Moore, Sebastian Riedel, and Pontus Stenetorp. 2022c. Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL), pages 8086-8098." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 750, + 524, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 750, + 524, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 750, + 524, + 772 + ], + "type": "text", + "content": "The mathlib Community. 2020. The lean mathematical library. In CPP 2020 - Proceedings of the 9th ACM" + } + ] + } + ], + "index": 20 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14619" + } + ] + } + ], + "index": 22 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 14 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 307, + 772 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 80, + 72, + 291, + 95 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 72, + 291, + 95 + ], + "spans": [ + { + "bbox": [ + 80, + 72, + 291, + 95 + ], + "type": "text", + "content": "SIGPLAN International Conference on Certified Programs and Proofs, co-located with POPL 2020." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 104, + 289, + 137 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 104, + 289, + 137 + ], + "spans": [ + { + "bbox": [ + 69, + 104, + 289, + 137 + ], + "type": "text", + "content": "Jordan Meadows and Andre Freitas. 2022. A survey in mathematical language processing. arXiv preprint arXiv:2205.15231." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 147, + 307, + 191 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 147, + 307, + 191 + ], + "spans": [ + { + "bbox": [ + 69, + 147, + 307, + 191 + ], + "type": "text", + "content": "Norman D. Megill and David A. Wheeler. 2019. Metamath: A Computer Language for Mathematical Proofs. Lulu Press, Morrisville, North Carolina. http://us.metamath.org/downloads/metamath.pdf." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 200, + 290, + 232 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 200, + 290, + 232 + ], + "spans": [ + { + "bbox": [ + 69, + 200, + 290, + 232 + ], + "type": "text", + "content": "Yuanliang Meng and Anna Rumshisky. 2019. Solving math word problems with double-decoder transformer. arXiv preprint arXiv:1908.10924." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 243, + 289, + 298 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 243, + 289, + 298 + ], + "spans": [ + { + "bbox": [ + 69, + 243, + 289, + 298 + ], + "type": "text", + "content": "Shen-Yun Miao, Chao-Chun Liang, and Keh-Yih Su. 2020. A diverse corpus for evaluating and developing english math word problem solvers. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 975-984." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 307, + 290, + 373 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 307, + 290, + 373 + ], + "spans": [ + { + "bbox": [ + 69, + 307, + 290, + 373 + ], + "type": "text", + "content": "Sewon Min, Xinxi Lyu, Ari Holtzman, Mikel Artetxe, Mike Lewis, Hannaneh Hajishirzi, and Luke Zettlemoyer. 2022. Rethinking the role of demonstrations: What makes in-context learning work? Proceedings of Empirical Methods in Natural Language Processing (EMNLP)." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 382, + 290, + 438 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 382, + 290, + 438 + ], + "spans": [ + { + "bbox": [ + 69, + 382, + 290, + 438 + ], + "type": "text", + "content": "Shervin Minaee, Nal Kalchbrenner, Erik Cambria, Narjes Nikzad, Meysam Chenaghlu, and Jianfeng Gao. 2021. Deep learning based text classification: a comprehensive review. ACM Computing Surveys (CSUR), 54(3):1-40." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 447, + 290, + 524 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 447, + 290, + 524 + ], + "spans": [ + { + "bbox": [ + 69, + 447, + 290, + 524 + ], + "type": "text", + "content": "Swaroop Mishra, Matthew Finlayson, Pan Lu, Leonard Tang, Sean Welleck, Chitta Baral, Tanmay Rajpurohit, Oyvind Tafjord, Ashish Sabharwal, Peter Clark, and Ashwin Kalyan. 2022a. Lila: A unified benchmark for mathematical reasoning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP)." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 534, + 290, + 609 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 534, + 290, + 609 + ], + "spans": [ + { + "bbox": [ + 69, + 534, + 290, + 609 + ], + "type": "text", + "content": "Swaroop Mishra, Arindam Mitra, Neeraj Varshney, Bhavdeep Sachdeva, Peter Clark, Chitta Baral, and Ashwin Kalyan. 2022b. Numglue: A suite of fundamental yet challenging mathematical reasoning tasks. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL), pages 3505-3523." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 619, + 290, + 708 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 619, + 290, + 708 + ], + "spans": [ + { + "bbox": [ + 69, + 619, + 290, + 708 + ], + "type": "text", + "content": "Eric Mitchell, Joseph J. Noh, Siyan Li, William S. Armstrong, Ananth Agarwal, Patrick Liu, Chelsea Finn, and Christopher D. Manning. 2022. Enhancing self-consistency and performance of pretrained language models with nli. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 717, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 717, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 717, + 290, + 772 + ], + "type": "text", + "content": "Leonardo de Moura, Soonho Kong, Jeremy Avigad, Floris van Doorn, and Jakob von Raumer. 2015. The lean theorem prover (system description). In International Conference on Automated Deduction, pages 378-388. Springer." + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 525, + 772 + ], + "type": "list", + "angle": 0, + "index": 23, + "blocks": [ + { + "bbox": [ + 305, + 72, + 525, + 137 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 305, + 72, + 525, + 137 + ], + "spans": [ + { + "bbox": [ + 305, + 72, + 525, + 137 + ], + "type": "text", + "content": "Reiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, et al. 2021. Webgpt: Browser-assisted question-answering with human feedback. arXiv preprint arXiv:2112.09332." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 147, + 525, + 200 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 147, + 525, + 200 + ], + "spans": [ + { + "bbox": [ + 304, + 147, + 525, + 200 + ], + "type": "text", + "content": "Allen Newell, John Clifford Shaw, and Herbert A Simon. 1957. Empirical explorations of the logic theory machine: A case study in heuristic. In Proceedings of the Western Joint Computer Conference, IRE-AIEE-ACM 1957." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 210, + 525, + 275 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 210, + 525, + 275 + ], + "spans": [ + { + "bbox": [ + 304, + 210, + 525, + 275 + ], + "type": "text", + "content": "Ansong Ni, Jeevana Priya Inala, Chenglong Wang, Oleksandr Polozov, Christopher Meek, Dragomir Radev, and Jianfeng Gao. 2023. Learning from self-sampled correct and partially-correct programs. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 285, + 525, + 350 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 285, + 525, + 350 + ], + "spans": [ + { + "bbox": [ + 304, + 285, + 525, + 350 + ], + "type": "text", + "content": "Maxwell Nye, Anders Johan Andreassen, Guy Gur-Ari, Henryk Michalewski, Jacob Austin, David Bieber, David Dohan, Aitor Lewkowycz, Maarten Bosma, David Luan, et al. 2021. Show your work: Scratchpads for intermediate computation with language models. arXiv preprint arXiv:2112.00114." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 359, + 525, + 424 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 359, + 525, + 424 + ], + "spans": [ + { + "bbox": [ + 304, + 359, + 525, + 424 + ], + "type": "text", + "content": "Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. 2022. Training language models to follow instructions with human feedback. In Advances in Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 433, + 525, + 509 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 433, + 525, + 509 + ], + "spans": [ + { + "bbox": [ + 304, + 433, + 525, + 509 + ], + "type": "text", + "content": "Arkil Patel, Satwik Bhattachamishra, and Navin Goyal. 2021. Are nlp models really able to solve simple math word problems? In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HIT), pages 2080-2094." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 518, + 525, + 562 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 518, + 525, + 562 + ], + "spans": [ + { + "bbox": [ + 304, + 518, + 525, + 562 + ], + "type": "text", + "content": "Lawrence C. Paulson. 1994. Isabelle - A Generic Theorem Prover (with a contribution by T. Nipkow), volume 828 of Lecture Notes in Computer Science. Springer." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 571, + 525, + 626 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 571, + 525, + 626 + ], + "spans": [ + { + "bbox": [ + 304, + 571, + 525, + 626 + ], + "type": "text", + "content": "Ethan Perez, Florian Strub, Harm De Vries, Vincent Dumoulin, and Aaron Courville. 2018. Film: Visual reasoning with a general conditioning layer. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 634, + 525, + 689 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 634, + 525, + 689 + ], + "spans": [ + { + "bbox": [ + 304, + 634, + 525, + 689 + ], + "type": "text", + "content": "Stanislas Polu, Jesse Michael Han, Kunhao Zheng, Mantas Baksys, Igor Babuschkin, and Ilya Sutskever. 2023. Formal mathematics statement curriculum learning. In International Conference on Learning Representations (ICLR), volume abs/2202.01344." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 698, + 525, + 731 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 698, + 525, + 731 + ], + "spans": [ + { + "bbox": [ + 304, + 698, + 525, + 731 + ], + "type": "text", + "content": "Stanislas Polu and Ilya Sutskever. 2020. Generative language modeling for automated theorem proving. arXiv preprint arXiv:2009.03393." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 304, + 739, + 525, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 739, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 739, + 525, + 772 + ], + "type": "text", + "content": "Jinghui Qin, Xiaodan Liang, Yining Hong, Jianheng Tang, and Liang Lin. 2021. Neural-symbolic solver for math word problems with auxiliary tasks. In" + } + ] + } + ], + "index": 22 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14620" + } + ] + } + ], + "index": 24 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 15 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 290, + 772 + ], + "type": "list", + "angle": 0, + "index": 11, + "blocks": [ + { + "bbox": [ + 80, + 72, + 290, + 117 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 72, + 290, + 117 + ], + "spans": [ + { + "bbox": [ + 80, + 72, + 290, + 117 + ], + "type": "text", + "content": "Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL), pages 5870-5881." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 124, + 289, + 190 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 124, + 289, + 190 + ], + "spans": [ + { + "bbox": [ + 69, + 124, + 289, + 190 + ], + "type": "text", + "content": "Jinghui Qin, Lihui Lin, Xiaodan Liang, Rumin Zhang, and Liang Lin. 2020. Semantically-aligned universal tree-structured solver for math word problems. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3780-3789." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 199, + 289, + 254 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 199, + 289, + 254 + ], + "spans": [ + { + "bbox": [ + 69, + 199, + 289, + 254 + ], + "type": "text", + "content": "Liang Qiu, Yizhou Zhao, Jinchao Li, Pan Lu, Baolin Peng, Jianfeng Gao, and Song-Chun Zhu. 2022a. Valuenet: A new dataset for human value driven dialogue system. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 2468-2484." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 262, + 289, + 328 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 262, + 289, + 328 + ], + "spans": [ + { + "bbox": [ + 69, + 262, + 289, + 328 + ], + "type": "text", + "content": "Liang Qiu, Yizhou Zhao, Yuan Liang, Pan Lu, Weiyan Shi, Zhou Yu, and Song-chun Zhu. 2022b. Towards socially intelligent agents with mental state transition and human value. In Proceedings of the 23rd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 146-158." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 337, + 289, + 391 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 337, + 289, + 391 + ], + "spans": [ + { + "bbox": [ + 69, + 337, + 289, + 391 + ], + "type": "text", + "content": "Xipeng Qiu, Tianxiang Sun, Yige Xu, Yunfan Shao, Ning Dai, and Xuanjing Huang. 2020. Pre-trained models for natural language processing: A survey. Science China Technological Sciences, 63(10):1872-1897." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 400, + 289, + 444 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 400, + 289, + 444 + ], + "spans": [ + { + "bbox": [ + 69, + 400, + 289, + 444 + ], + "type": "text", + "content": "Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. 2020. Language models are unsupervised multitask learners. OpenAI Blog." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 452, + 289, + 518 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 452, + 289, + 518 + ], + "spans": [ + { + "bbox": [ + 69, + 452, + 289, + 518 + ], + "type": "text", + "content": "Jack W Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young, et al. 2021. Scaling language models: Methods, analysis & insights from training gopher. arXiv preprint arXiv:2112.11446." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 527, + 289, + 592 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 527, + 289, + 592 + ], + "spans": [ + { + "bbox": [ + 69, + 527, + 289, + 592 + ], + "type": "text", + "content": "Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research (JMLR), 21:1-67." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 601, + 289, + 667 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 601, + 289, + 667 + ], + "spans": [ + { + "bbox": [ + 69, + 601, + 289, + 667 + ], + "type": "text", + "content": "Abhilasha Ravichander, Aakanksha Naik, Carolyn Rose, and Eduard Hovy. 2019. Equate: A benchmark evaluation framework for quantitative reasoning in natural language inference. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 349-361." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 675, + 289, + 731 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 675, + 289, + 731 + ], + "spans": [ + { + "bbox": [ + 69, + 675, + 289, + 731 + ], + "type": "text", + "content": "Yasaman Razeghi, Robert L Logan IV, Matt Gardner, and Sameer Singh. 2022. Impact of pretraining term frequencies on few-shot numerical reasoning. In *Findings of the Association for Computational Linguistics: EMNLP* 2022, pages 840-854." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 739, + 289, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 739, + 289, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 739, + 289, + 772 + ], + "type": "text", + "content": "Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. Advances" + } + ] + } + ], + "index": 10 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 524, + 772 + ], + "type": "list", + "angle": 0, + "index": 25, + "blocks": [ + { + "bbox": [ + 315, + 72, + 524, + 94 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 315, + 72, + 524, + 94 + ], + "spans": [ + { + "bbox": [ + 315, + 72, + 524, + 94 + ], + "type": "text", + "content": "in neural information processing systems (NeurIPS), 28." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 103, + 524, + 158 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 103, + 524, + 158 + ], + "spans": [ + { + "bbox": [ + 304, + 103, + 524, + 158 + ], + "type": "text", + "content": "Ryokan Ri and Yoshimasa Tsuruoka. 2022. Pretraining with artificial language: Studying transferable knowledge in language models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL), pages 7302-7315." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 167, + 524, + 211 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 167, + 524, + 211 + ], + "spans": [ + { + "bbox": [ + 304, + 167, + 524, + 211 + ], + "type": "text", + "content": "Benjamin Robaidek, Rik Koncel-Kedziorski, and Hannaneh Hajishirzi. 2018. Data-driven methods for solving algebra word problems. arXiv preprint arXiv:1804.10718." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 219, + 524, + 264 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 219, + 524, + 264 + ], + "spans": [ + { + "bbox": [ + 304, + 219, + 524, + 264 + ], + "type": "text", + "content": "Subhro Roy and Dan Roth. 2015. Solving general arithmetic word problems. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1743-1752." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 272, + 524, + 317 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 272, + 524, + 317 + ], + "spans": [ + { + "bbox": [ + 304, + 272, + 524, + 317 + ], + "type": "text", + "content": "Subhro Roy and Dan Roth. 2017. Unit dependency graph and its application to arithmetic word problem solving. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 325, + 524, + 370 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 325, + 524, + 370 + ], + "spans": [ + { + "bbox": [ + 304, + 325, + 524, + 370 + ], + "type": "text", + "content": "Subhro Roy and Dan Roth. 2018. Mapping to declarative knowledge for word problem solving. Transactions of the Association for Computational Linguistics (TACL), 6:159-172." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 378, + 524, + 422 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 378, + 524, + 422 + ], + "spans": [ + { + "bbox": [ + 304, + 378, + 524, + 422 + ], + "type": "text", + "content": "Subhro Roy, Tim Vieira, and Dan Roth. 2015. Reasoning about quantities in natural language. Transactions of the Association for Computational Linguistics (TACL), 3:1-13." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 431, + 524, + 476 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 431, + 524, + 476 + ], + "spans": [ + { + "bbox": [ + 304, + 431, + 524, + 476 + ], + "type": "text", + "content": "Ohad Rubin, Jonathan Herzig, and Jonathan Berant. 2022. Learning to retrieve prompts for in-context learning. North American Chapter of the Association for Computational Linguistics (NAACL)." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 483, + 524, + 550 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 483, + 524, + 550 + ], + "spans": [ + { + "bbox": [ + 304, + 483, + 524, + 550 + ], + "type": "text", + "content": "Mrinmaya Sachan, Kumar Dubey, and Eric Xing. 2017. From textbooks to knowledge: A case study in harvesting axiomatic knowledge from textbooks to solve geometry problems. In Proceedings of Empirical Methods in Natural Language Processing (EMNLP), pages 773-784." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 558, + 524, + 613 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 558, + 524, + 613 + ], + "spans": [ + { + "bbox": [ + 304, + 558, + 524, + 613 + ], + "type": "text", + "content": "Mrinmaya Sachan and Eric Xing. 2017. Learning to solve geometry problems from natural language demonstrations in textbooks. In Proceedings of the 6th Joint Conference on Lexical and Computational Semantics, pages 251-261." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 304, + 622, + 524, + 666 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 622, + 524, + 666 + ], + "spans": [ + { + "bbox": [ + 304, + 622, + 524, + 666 + ], + "type": "text", + "content": "David Saxton, Edward Grefenstette, Felix Hill, and Pushmeet Kohli. 2020. Analysing mathematical reasoning abilities of neural models. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 304, + 675, + 524, + 729 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 675, + 524, + 729 + ], + "spans": [ + { + "bbox": [ + 304, + 675, + 524, + 729 + ], + "type": "text", + "content": "Tal Schuster, Ashwin Kalyan, Alex Polozov, and Adam Tauman Kalai. 2021. Programming puzzles. In Thirty-fifth Conference on Neural Information Processing Systems (NeurIPS) Datasets and Benchmarks Track." + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 304, + 739, + 524, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 739, + 524, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 739, + 524, + 772 + ], + "type": "text", + "content": "Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual" + } + ] + } + ], + "index": 24 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 311, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 311, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 311, + 791 + ], + "type": "text", + "content": "14621" + } + ] + } + ], + "index": 26 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 16 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 772 + ], + "type": "list", + "angle": 0, + "index": 12, + "blocks": [ + { + "bbox": [ + 80, + 72, + 291, + 95 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 72, + 291, + 95 + ], + "spans": [ + { + "bbox": [ + 80, + 72, + 291, + 95 + ], + "type": "text", + "content": "Meeting of the Association for Computational Linguistics (ACL), pages 1715-1725." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 103, + 289, + 158 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 103, + 289, + 158 + ], + "spans": [ + { + "bbox": [ + 69, + 103, + 289, + 158 + ], + "type": "text", + "content": "Minjoon Seo, Hannaneh Hajishirzi, Ali Farhadi, Oren Etzioni, and Clint Malcolm. 2015. Solving geometry problems: Combining text and diagram interpretation. In Proceedings of Empirical Methods in Natural Language Processing (EMNLP), pages 1466-1476." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 167, + 289, + 222 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 167, + 289, + 222 + ], + "spans": [ + { + "bbox": [ + 69, + 167, + 289, + 222 + ], + "type": "text", + "content": "Jianhao Shen, Yichun Yin, Lin Li, Lifeng Shang, Xin Jiang, Ming Zhang, and Qun Liu. 2021. Generate & rank: A multi-task framework for math word problems. In Findings of the Association for Computational Linguistics (EMNLP), pages 2269-2279." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 230, + 289, + 284 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 230, + 289, + 284 + ], + "spans": [ + { + "bbox": [ + 69, + 230, + 289, + 284 + ], + "type": "text", + "content": "Yibin Shen and Cheqing Jin. 2020. Solving math word problems with multi-encoders and multi-decoders. In Proceedings of the 28th International Conference on Computational Linguistics (COLING), pages 2924-2934." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 294, + 289, + 359 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 294, + 289, + 359 + ], + "spans": [ + { + "bbox": [ + 69, + 294, + 289, + 359 + ], + "type": "text", + "content": "Shuming Shi, Yuehui Wang, Chin-Yew Lin, Xiaojiang Liu, and Yong Rui. 2015. Automatically solving number word problems by semantic parsing and reasoning. In Proceedings of the 2015 conference on empirical methods in natural language processing (EMNLP), pages 1132-1142." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 369, + 289, + 412 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 369, + 289, + 412 + ], + "spans": [ + { + "bbox": [ + 69, + 369, + 289, + 412 + ], + "type": "text", + "content": "Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and TieYan Liu. 2019. Mass: Masked sequence to sequence pre-training for language generation. In 36th International Conference on Machine Learning (ICML)." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 421, + 289, + 476 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 421, + 289, + 476 + ], + "spans": [ + { + "bbox": [ + 69, + 421, + 289, + 476 + ], + "type": "text", + "content": "Kai Sun, Dian Yu, Jianshu Chen, Dong Yu, Yejin Choi, and Claire Cardie. 2019. Dream: A challenge data set and models for dialogue-based reading comprehension. Transactions of the Association for Computational Linguistics (TACL), 7:217-231." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 485, + 289, + 528 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 485, + 289, + 528 + ], + "spans": [ + { + "bbox": [ + 69, + 485, + 289, + 528 + ], + "type": "text", + "content": "Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems (NeurIPS), 27." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 537, + 289, + 592 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 537, + 289, + 592 + ], + "spans": [ + { + "bbox": [ + 69, + 537, + 289, + 592 + ], + "type": "text", + "content": "Oyvind Tafjord, Peter Clark, Matt Gardner, Wen-tau Yih, and Ashish Sabharwal. 2019. Quarel: A dataset and models for answering questions about qualitative relationships. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 7063-7071." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 601, + 289, + 677 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 601, + 289, + 677 + ], + "spans": [ + { + "bbox": [ + 69, + 601, + 289, + 677 + ], + "type": "text", + "content": "Kai Sheng Tai, Richard Socher, and Christopher D Manning. 2015. Improved semantic representations from tree-structured long short-term memory networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (ACL), pages 1556-1566." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 686, + 289, + 730 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 686, + 289, + 730 + ], + "spans": [ + { + "bbox": [ + 69, + 686, + 289, + 730 + ], + "type": "text", + "content": "Avijit Thawani, Jay Pujara, and Ashwin Kalyan. 2022. Estimating numbers without regression. In 36th Conference on Neural Information Processing Systems (NeurIPS 2022) Workshop on MATH-AI." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 69, + 739, + 289, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 739, + 289, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 739, + 289, + 772 + ], + "type": "text", + "content": "Avijit Thawani, Jay Pujara, Pedro A Szekely, and Filip Ilievski. 2021. Representing numbers in nlp: a survey and a vision. In Proceedings of the 2021 Conference" + } + ] + } + ], + "index": 11 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 525, + 772 + ], + "type": "list", + "angle": 0, + "index": 24, + "blocks": [ + { + "bbox": [ + 314, + 72, + 524, + 105 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 72, + 524, + 105 + ], + "spans": [ + { + "bbox": [ + 314, + 72, + 524, + 105 + ], + "type": "text", + "content": "of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HIT), pages 644-656." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 117, + 525, + 172 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 117, + 525, + 172 + ], + "spans": [ + { + "bbox": [ + 304, + 117, + 525, + 172 + ], + "type": "text", + "content": "Shounaak Ughade and Satish Kumbhar. 2019. Survey on mathematical word problem solving using natural language processing. In 2019 1st International Conference on Innovations in Information and Communication Technology (ICIICT), pages 1-5. IEEE." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 184, + 525, + 216 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 184, + 525, + 216 + ], + "spans": [ + { + "bbox": [ + 304, + 184, + 525, + 216 + ], + "type": "text", + "content": "Shyam Upadhyay and Ming-Wei Chang. 2015. Draw: A challenging and diverse algebra word problem set. Technical report, Citeseer." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 228, + 525, + 294 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 228, + 525, + 294 + ], + "spans": [ + { + "bbox": [ + 304, + 228, + 525, + 294 + ], + "type": "text", + "content": "Shyam Upadhyay and Ming-Wei Chang. 2017. Annotating derivations: A new evaluation strategy and dataset for algebra word problems. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics (ACL), pages 494-504." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 306, + 525, + 338 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 306, + 525, + 338 + ], + "spans": [ + { + "bbox": [ + 304, + 306, + 525, + 338 + ], + "type": "text", + "content": "Josef Urban. 2006. Mptp 0.2: Design, implementation, and initial experiments. Journal of Automated Reasoning, 37(1):21-43." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 350, + 525, + 406 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 350, + 525, + 406 + ], + "spans": [ + { + "bbox": [ + 304, + 350, + 525, + 406 + ], + "type": "text", + "content": "Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems (NeurIPS), pages 5998-6008." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 417, + 525, + 494 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 417, + 525, + 494 + ], + "spans": [ + { + "bbox": [ + 304, + 417, + 525, + 494 + ], + "type": "text", + "content": "Eric Wallace, Yizhong Wang, Sujian Li, Sameer Singh, and Matt Gardner. 2019. Do nlp models know numbers? probing numeracy in embeddings. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5307-5315." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 506, + 525, + 561 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 506, + 525, + 561 + ], + "spans": [ + { + "bbox": [ + 304, + 506, + 525, + 561 + ], + "type": "text", + "content": "Lei Wang, Yan Wang, Deng Cai, Dongxiang Zhang, and Xiaojiang Liu. 2018a. Translating a math word problem to a expression tree. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1064-1069." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 572, + 525, + 628 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 572, + 525, + 628 + ], + "spans": [ + { + "bbox": [ + 304, + 572, + 525, + 628 + ], + "type": "text", + "content": "Lei Wang, Dongxiang Zhang, Lianli Gao, Jingkuan Song, Long Guo, and Heng Tao Shen. 2018b. Math-dqn: Solving arithmetic word problems via deep reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 304, + 640, + 525, + 704 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 640, + 525, + 704 + ], + "spans": [ + { + "bbox": [ + 304, + 640, + 525, + 704 + ], + "type": "text", + "content": "Lei Wang, Dongxiang Zhang, Jipeng Zhang, Xing Xu, Lianli Gao, Bing Tian Dai, and Heng Tao Shen. 2019. Template-based math word problem solvers with recursive neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 7144-7151." + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 304, + 717, + 525, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 717, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 717, + 525, + 772 + ], + "type": "text", + "content": "Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc Le, Ed Chi, and Denny Zhou. 2023. Self-consistency improves chain of thought reasoning in language models. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 23 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14622" + } + ] + } + ], + "index": 25 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 17 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 772 + ], + "type": "list", + "angle": 0, + "index": 12, + "blocks": [ + { + "bbox": [ + 69, + 72, + 291, + 128 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 72, + 291, + 128 + ], + "spans": [ + { + "bbox": [ + 69, + 72, + 291, + 128 + ], + "type": "text", + "content": "Yan Wang, Xiaojiang Liu, and Shuming Shi. 2017. Deep neural solver for math word problems. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 845-854." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 136, + 290, + 191 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 136, + 290, + 191 + ], + "spans": [ + { + "bbox": [ + 69, + 136, + 290, + 191 + ], + "type": "text", + "content": "Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Ed Chi, Quoc Le, and Denny Zhou. 2022. Chain of thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 200, + 290, + 265 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 200, + 290, + 265 + ], + "spans": [ + { + "bbox": [ + 69, + 200, + 290, + 265 + ], + "type": "text", + "content": "Sean Welleck, Jiacheng Liu, Ronan Le Bras, Hannaneh Hajishirzi, Yejin Choi, and Kyunghyun Cho. 2021. Naturalproofs: Mathematical theorem proving in natural language. In Thirty-fifth Conference on Neural Information Processing Systems (NeurIPS) Datasets and Benchmarks Track." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 274, + 290, + 329 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 274, + 290, + 329 + ], + "spans": [ + { + "bbox": [ + 69, + 274, + 290, + 329 + ], + "type": "text", + "content": "Sean Welleck, Jiacheng Liu, Ximing Lu, Hannaneh Hajishirzi, and Yejin Choi. 2022a. Naturalprover: Grounded mathematical proof generation with language models. In Advances in Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 338, + 290, + 393 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 338, + 290, + 393 + ], + "spans": [ + { + "bbox": [ + 69, + 338, + 290, + 393 + ], + "type": "text", + "content": "Sean Welleck, Ximing Lu, Peter West, Faeze Brahman, Tianxiao Shen, Daniel Khashabi, and Yejin Choi. 2023. Generating sequences by learning to self-correct. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 401, + 290, + 444 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 401, + 290, + 444 + ], + "spans": [ + { + "bbox": [ + 69, + 401, + 290, + 444 + ], + "type": "text", + "content": "Sean Welleck, Peter West, Jize Cao, and Yejin Choi. 2022b. Symbolic brittleness in sequence models: on systematic generalization in symbolic mathematics. In AAAI." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 453, + 290, + 487 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 453, + 290, + 487 + ], + "spans": [ + { + "bbox": [ + 69, + 453, + 290, + 487 + ], + "type": "text", + "content": "Wu Wen-Tsun. 1986. Basic principles of mechanical theorem proving in elementary geometries. Journal of automated Reasoning, 2(3):221-252." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 495, + 290, + 528 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 495, + 290, + 528 + ], + "spans": [ + { + "bbox": [ + 69, + 495, + 290, + 528 + ], + "type": "text", + "content": "Daniel Whalen. 2016. Holophrasm: a neural automated theorem prover for higher-order logic. arXiv preprint arXiv:1608.02644." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 537, + 290, + 582 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 537, + 290, + 582 + ], + "spans": [ + { + "bbox": [ + 69, + 537, + 290, + 582 + ], + "type": "text", + "content": "Sanghyun Woo, Jongchan Park, Joon-Young Lee, and In So Kweon. 2018. Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV), pages 3-19." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 590, + 290, + 655 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 590, + 290, + 655 + ], + "spans": [ + { + "bbox": [ + 69, + 590, + 290, + 655 + ], + "type": "text", + "content": "Qinzhuo Wu, Qi Zhang, Jinlan Fu, and Xuan-Jing Huang. 2020. A knowledge-aware sequence-to-tree network for math word problem solving. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7137-7146." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 664, + 290, + 719 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 664, + 290, + 719 + ], + "spans": [ + { + "bbox": [ + 69, + 664, + 290, + 719 + ], + "type": "text", + "content": "Qinzhuo Wu, Qi Zhang, and Zhongyu Wei. 2021a. An edge-enhanced hierarchical graph-to-tree network for math word problem solving. In *Findings of the Association for Computational Linguistics (EMNLP)*, pages 1473-1482." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 69, + 728, + 290, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 728, + 290, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 728, + 290, + 772 + ], + "type": "text", + "content": "Qinzhuo Wu, Qi Zhang, Zhongyu Wei, and Xuan-Jing Huang. 2021b. Math word problem solving with explicit numerical values. In Proceedings of the 59th Annual Meeting of the Association for Computational" + } + ] + } + ], + "index": 11 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 524, + 772 + ], + "type": "list", + "angle": 0, + "index": 25, + "blocks": [ + { + "bbox": [ + 314, + 72, + 524, + 105 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 72, + 524, + 105 + ], + "spans": [ + { + "bbox": [ + 314, + 72, + 524, + 105 + ], + "type": "text", + "content": "Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL), pages 5859-5869." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 112, + 524, + 158 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 112, + 524, + 158 + ], + "spans": [ + { + "bbox": [ + 304, + 112, + 524, + 158 + ], + "type": "text", + "content": "Xingjiao Wu, Luwei Xiao, Yixuan Sun, Junhang Zhang, Tianlong Ma, and Liang He. 2022a. A survey of human-in-the-loop for machine learning. Future Generation Computer Systems." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 164, + 524, + 231 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 164, + 524, + 231 + ], + "spans": [ + { + "bbox": [ + 304, + 164, + 524, + 231 + ], + "type": "text", + "content": "Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. 2016. Google's neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 238, + 524, + 293 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 238, + 524, + 293 + ], + "spans": [ + { + "bbox": [ + 304, + 238, + 524, + 293 + ], + "type": "text", + "content": "Yuhuai Wu, Albert Jiang, Jimmy Ba, and Roger Baker Grosse. 2021c. Int: An inequality benchmark for evaluating generalization in theorem proving. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 301, + 524, + 356 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 301, + 524, + 356 + ], + "spans": [ + { + "bbox": [ + 304, + 301, + 524, + 356 + ], + "type": "text", + "content": "Yuhuai Wu, Albert Qiaochu Jiang, Wenda Li, Markus Norman Rabe, Charles E Staats, Mateja Jamnik, and Christian Szegedy. 2022b. Autoformalization with large language models. In Advances in Neural Information Processing Systems (NeurIPS)." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 363, + 524, + 396 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 363, + 524, + 396 + ], + "spans": [ + { + "bbox": [ + 304, + 363, + 524, + 396 + ], + "type": "text", + "content": "Yuhuai Wu, Felix Li, and Percy Liang. 2022c. Insights into pre-training via simpler synthetic tasks. arXiv preprint arXiv:2206.10139." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 403, + 524, + 470 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 403, + 524, + 470 + ], + "spans": [ + { + "bbox": [ + 304, + 403, + 524, + 470 + ], + "type": "text", + "content": "Yuhuai Wu, Markus N Rabe, Wenda Li, Jimmy Ba, Roger B Grosse, and Christian Szegedy. 2021d. Lime: Learning inductive bias for primitives of mathematical reasoning. In International Conference on Machine Learning (ICML), pages 11251-11262. PMLR." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 477, + 524, + 522 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 477, + 524, + 522 + ], + "spans": [ + { + "bbox": [ + 304, + 477, + 524, + 522 + ], + "type": "text", + "content": "Zhipeng Xie and Shichao Sun. 2019. A goal-driven tree-structured neural model for math word problems. In International Joint Conference on Artificial Intelligence (IJCAI), pages 5299-5305." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 528, + 524, + 596 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 528, + 524, + 596 + ], + "spans": [ + { + "bbox": [ + 304, + 528, + 524, + 596 + ], + "type": "text", + "content": "Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. 2015. Show, attend and tell: Neural image caption generation with visual attention. In International conference on machine learning (ICML), pages 2048-2057. PMLR." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 304, + 602, + 524, + 647 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 602, + 524, + 647 + ], + "spans": [ + { + "bbox": [ + 304, + 602, + 524, + 647 + ], + "type": "text", + "content": "Kaiyu Yang and Jia Deng. 2019. Learning to prove theorems via interacting with proof assistants. In International Conference on Machine Learning (ICML), pages 6984-6994. PMLR." + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 304, + 654, + 524, + 699 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 654, + 524, + 699 + ], + "spans": [ + { + "bbox": [ + 304, + 654, + 524, + 699 + ], + "type": "text", + "content": "Zheng Ye, Shang-Ching Chou, and Xiao-Shan Gao. 2008. An introduction to java geometry expert. In International workshop on automated deduction in geometry, pages 189-195. Springer." + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 304, + 706, + 524, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 706, + 524, + 772 + ], + "spans": [ + { + "bbox": [ + 304, + 706, + 524, + 772 + ], + "type": "text", + "content": "Wei Yu, Mengzhu Wang, Xiaodong Wang, Xun Zhou, Yongfu Zha, Yongjian Zhang, Shuyu Miao, and Jingdong Liu. 2021a. Geore: A relation extraction dataset for Chinese geometry problems. In 35th Conference on Neural Information Processing Systems (NeurIPS) Workshop on Math AI for Education (MATHAI4ED)." + } + ] + } + ], + "index": 24 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14623" + } + ] + } + ], + "index": 26 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 18 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 72, + 289, + 772 + ], + "type": "list", + "angle": 0, + "index": 10, + "blocks": [ + { + "bbox": [ + 69, + 72, + 289, + 138 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 72, + 289, + 138 + ], + "spans": [ + { + "bbox": [ + 69, + 72, + 289, + 138 + ], + "type": "text", + "content": "Weijiang Yu, Yingpeng Wen, Fudan Zheng, and Nong Xiao. 2021b. Improving math word problems with pre-trained knowledge and hierarchical reasoning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3384-3394." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 149, + 289, + 216 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 149, + 289, + 216 + ], + "spans": [ + { + "bbox": [ + 69, + 149, + 289, + 216 + ], + "type": "text", + "content": "Wenhao Yu, Dan Iter, Shuohang Wang, Yichong Xu, Mingxuan Ju, Soumya Sanyal, Chenguang Zhu, Michael Zeng, and Meng Jiang. 2023. Generate rather than retrieve: Large language models are strong context generators. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 227, + 289, + 282 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 227, + 289, + 282 + ], + "spans": [ + { + "bbox": [ + 69, + 227, + 289, + 282 + ], + "type": "text", + "content": "Klim Zaporojets, Giannis Bekoulis, Johannes Deleu, Thomas Demeester, and Chris Develder. 2021. Solving arithmetic word problems by scoring equations with recursive neural networks. Expert Systems with Applications, 174:114704." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 294, + 289, + 349 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 294, + 289, + 349 + ], + "spans": [ + { + "bbox": [ + 69, + 294, + 289, + 349 + ], + "type": "text", + "content": "Dongxiang Zhang, Lei Wang, Luming Zhang, Bing Tian Dai, and Heng Tao Shen. 2019. The gap of semantic parsing: A survey on automatic math word problem solvers. IEEE transactions on pattern analysis and machine intelligence, 42(9):2287-2305." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 361, + 289, + 416 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 361, + 289, + 416 + ], + "spans": [ + { + "bbox": [ + 69, + 361, + 289, + 416 + ], + "type": "text", + "content": "Jipeng Zhang, Roy Ka-Wei Lee, Ee-Peng Lim, Wei Qin, Lei Wang, Jie Shao, and Qianru Sun. 2020a. Teacher-student networks with multiple decoders for solving math word problem. In International Joint Conference on Artificial Intelligence (IJCAI)." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 428, + 289, + 492 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 428, + 289, + 492 + ], + "spans": [ + { + "bbox": [ + 69, + 428, + 289, + 492 + ], + "type": "text", + "content": "Jipeng Zhang, Lei Wang, Roy Ka-Wei Lee, Yi Bin, Yan Wang, Jie Shao, and Ee-Peng Lim. 2020b. Graph-to-tree learning for solving math word problems. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 3928-3937." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 506, + 289, + 560 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 506, + 289, + 560 + ], + "spans": [ + { + "bbox": [ + 69, + 506, + 289, + 560 + ], + "type": "text", + "content": "Ming-Liang Zhang, Fei Yin, Yi-Han Hao, and ChengLin Liu. 2022. Learning to understand plane geometry diagram. In 36th Conference on Neural Information Processing Systems (NeurIPS) Workshop on MATH-AI." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 572, + 289, + 638 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 572, + 289, + 638 + ], + "spans": [ + { + "bbox": [ + 69, + 572, + 289, + 638 + ], + "type": "text", + "content": "Qiyuan Zhang, Lei Wang, Sicheng Yu, Shuohang Wang, Yang Wang, Jing Jiang, and Ee-Peng Lim. 2021. Noahqa: Numerical reasoning with interpretable graph question answering dataset. In Findings of the Association for Computational Linguistics (EMNLP), pages 4147-4161." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 650, + 289, + 705 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 650, + 289, + 705 + ], + "spans": [ + { + "bbox": [ + 69, + 650, + 289, + 705 + ], + "type": "text", + "content": "Wenhe Zhang, Chi Zhang, Yixin Zhu, and Song-Chun Zhu. 2020c. Machine number sense: A dataset of visual arithmetic problems for abstract and relational reasoning. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 1332-1340." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 717, + 289, + 772 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 717, + 289, + 772 + ], + "spans": [ + { + "bbox": [ + 69, + 717, + 289, + 772 + ], + "type": "text", + "content": "Xikun Zhang, Deepak Ramachandran, Ian Tenney, Yanai Elazar, and Dan Roth. 2020d. Do language embeddings capture scales? In Proceedings of the Third Blackbox NLP Workshop on Analyzing and Interpreting Neural Networks for NLP, pages 292-299." + } + ] + } + ], + "index": 9 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 524, + 756 + ], + "type": "list", + "angle": 0, + "index": 21, + "blocks": [ + { + "bbox": [ + 304, + 72, + 524, + 127 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 72, + 524, + 127 + ], + "spans": [ + { + "bbox": [ + 304, + 72, + 524, + 127 + ], + "type": "text", + "content": "Xikun Zhang, Deepak Ramachandran, Ian Tenney, Yanai Elazar, and Dan Roth. 2020e. Do language embeddings capture scales? In Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, pages 292-299." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 304, + 136, + 524, + 212 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 136, + 524, + 212 + ], + "spans": [ + { + "bbox": [ + 304, + 136, + 524, + 212 + ], + "type": "text", + "content": "Yizhe Zhang, Siqi Sun, Michel Galley, Yen-Chun Chen, Chris Brockett, Xiang Gao, Jianfeng Gao, Jingjing Liu, and Bill Dolan. 2020f. Dialogpt: Large-scale generative pre-training for conversational response generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 304, + 222, + 524, + 265 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 222, + 524, + 265 + ], + "spans": [ + { + "bbox": [ + 304, + 222, + 524, + 265 + ], + "type": "text", + "content": "Zhuosheng Zhang, Aston Zhang, Mu Li, and Alex Smola. 2023. Automatic chain of thought prompting in large language models. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 304, + 275, + 524, + 318 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 275, + 524, + 318 + ], + "spans": [ + { + "bbox": [ + 304, + 275, + 524, + 318 + ], + "type": "text", + "content": "Wei Zhao, Mingyue Shang, Yang Liu, Liang Wang, and Jingming Liu. 2020. Ape210k: A large-scale and template-rich dataset of math word problems. arXiv preprint arXiv:2009.11506." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 327, + 524, + 382 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 327, + 524, + 382 + ], + "spans": [ + { + "bbox": [ + 304, + 327, + 524, + 382 + ], + "type": "text", + "content": "Yilun Zhao, Yunxiang Li, Chenying Li, and Rui Zhang. 2022. Multihiertt: Numerical reasoning over multi hierarchical tabular and textual data. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL), pages 6588-6600." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 391, + 524, + 445 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 391, + 524, + 445 + ], + "spans": [ + { + "bbox": [ + 304, + 391, + 524, + 445 + ], + "type": "text", + "content": "Zihao Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate before use: Improving few-shot performance of language models. In International Conference on Machine Learning (ICML), pages 12697-12706. PMLR." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 454, + 524, + 498 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 454, + 524, + 498 + ], + "spans": [ + { + "bbox": [ + 304, + 454, + 524, + 498 + ], + "type": "text", + "content": "Kunhao Zheng, Jesse Michael Han, and Stanislas Polu. 2022. Minif2f: a cross-system benchmark for formal olympiad-level mathematics. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 507, + 524, + 572 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 507, + 524, + 572 + ], + "spans": [ + { + "bbox": [ + 304, + 507, + 524, + 572 + ], + "type": "text", + "content": "Ben Zhou, Daniel Khashabi, Qiang Ning, and Dan Roth. 2019. \"Going on a vacation\" takes longer than \"Going for a walk\": A Study of Temporal Common-sense Understanding. In Proc. of the Conference on Empirical Methods in Natural Language Processing (EMNLP)." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 581, + 524, + 647 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 581, + 524, + 647 + ], + "spans": [ + { + "bbox": [ + 304, + 581, + 524, + 647 + ], + "type": "text", + "content": "Denny Zhou, Nathanael Scharli, Le Hou, Jason Wei, Nathan Scales, Xuezhi Wang, Dale Schuurmans, Olivier Bousquet, Quoc Le, and Ed Chi. 2023. Least-to-most prompting enables complex reasoning in large language models. In International Conference on Learning Representations (ICLR)." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 656, + 524, + 756 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 656, + 524, + 756 + ], + "spans": [ + { + "bbox": [ + 304, + 656, + 524, + 756 + ], + "type": "text", + "content": "Fengbin Zhu, Wenqiang Lei, Youcheng Huang, Chao Wang, Shuo Zhang, Jiancheng Lv, Fuli Feng, and Tat-Seng Chua. 2021. Tat-qa: A question answering benchmark on a hybrid of tabular and textual content in finance. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL-JCNLP), pages 3277-3287." + } + ] + } + ], + "index": 20 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14624" + } + ] + } + ], + "index": 22 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 19 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 69, + 69, + 287, + 191 + ], + "blocks": [ + { + "bbox": [ + 69, + 69, + 287, + 191 + ], + "lines": [ + { + "bbox": [ + 69, + 69, + 287, + 191 + ], + "spans": [ + { + "bbox": [ + 69, + 69, + 287, + 191 + ], + "type": "image", + "image_path": "9a06341e53ae457db25b42a9f760bd3755339670ed4290d411539c38b8d0477e.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 67, + 197, + 290, + 234 + ], + "lines": [ + { + "bbox": [ + 67, + 197, + 290, + 234 + ], + "spans": [ + { + "bbox": [ + 67, + 197, + 290, + 234 + ], + "type": "text", + "content": "Figure 3: Estimated counts of annually published papers on deep learning for mathematical reasoning. This field has been experiencing rapid growth since 2018." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 68, + 245, + 265, + 259 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 245, + 265, + 259 + ], + "spans": [ + { + "bbox": [ + 68, + 245, + 265, + 259 + ], + "type": "text", + "content": "A Mathematical Reasoning Datasets" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 268, + 290, + 335 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 268, + 290, + 335 + ], + "spans": [ + { + "bbox": [ + 67, + 268, + 290, + 335 + ], + "type": "text", + "content": "In this section, we will examine the various datasets currently available for the study of mathematical reasoning using deep learning methods. A summary of the commonly used datasets in this field can be found in Table 7." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 68, + 347, + 233, + 360 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 347, + 233, + 360 + ], + "spans": [ + { + "bbox": [ + 68, + 347, + 233, + 360 + ], + "type": "text", + "content": "A.1 Math Word Problem Solving" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 366, + 291, + 582 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 366, + 291, + 582 + ], + "spans": [ + { + "bbox": [ + 67, + 366, + 291, + 582 + ], + "type": "text", + "content": "Developing algorithms to solve math word problems (MwPs) automatically has been an interest of NLP researchers for decades (Feigenbaum et al., 1963; Bobrow, 1964). A math word problem (also termed an algebraic or arithmetic word problem) describes a brief narrative that involves characters, entities, and quantities. The mathematical relationship of an MwP can be modeled with a set of equations whose solution reveals the final answer to the question. A typical example is shown in Table 1. A question involves the four basic arithmetic operations of addition, subtraction, multiplication, and division with single or multiple operation steps. The challenge of MwPs for NLP systems lies in the need for language comprehension, semantic parsing, and multiple mathematical reasoning skills." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 584, + 291, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 584, + 291, + 772 + ], + "spans": [ + { + "bbox": [ + 67, + 584, + 291, + 772 + ], + "type": "text", + "content": "Existing MWP datasets cover grade school problems, which are crawled from online learning websites (Koncel-Kedziorski et al., 2015), collected from textbooks, or manually annotated by human workers (Patel et al., 2021). Early math word problem datasets are relatively small or limited to a small number of operation steps (Hosseini et al., 2014; Kushman et al., 2014; Roy et al., 2015). Some recently curated datasets aim to increase problem diversity and difficulty levels. For example, Ape210K (Zhao et al., 2020) consists of 210k elementary math word problems, which is the largest publicly available. The problems in GSM8K (Cobbe et al., 2021) can involve up to 8 steps to" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 71, + 526, + 179 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 526, + 179 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 526, + 179 + ], + "type": "text", + "content": "solve. SVAMP (Patel et al., 2021) is a benchmark that tests the robustness of deep learning models to math word problems with simple variations. More recently built datasets involve modalities beyond text. For example, IconQA (Lu et al., 2021b) provides an abstract diagram as a visual context, while TabMWP (Lu et al., 2022b) provides a tabular context for each problem." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 181, + 526, + 357 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 181, + 526, + 357 + ], + "spans": [ + { + "bbox": [ + 302, + 181, + 526, + 357 + ], + "type": "text", + "content": "Most MWP datasets provide annotated equations as a rationale for the solution (e.g., Table 1). To improve the performance and interpretability of the learned solvers, MathQA (Tafjord et al., 2019) is annotated with precise operation programs, and MathQA-Python (Austin et al., 2021) is provided with specific Python programs instead. Another line of datasets annotates the problems with multi-step natural language solutions that are regarded as more human-readable (Ling et al., 2017; Cobbe et al., 2021; Lu et al., 2022b). Lila (Mishra et al., 2022a) annotates many of the previously mentioned MWP datasets with Python program rationales." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 303, + 371, + 415, + 384 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 371, + 415, + 384 + ], + "spans": [ + { + "bbox": [ + 303, + 371, + 415, + 384 + ], + "type": "text", + "content": "A.2 Theorem Proving" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 392, + 525, + 568 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 392, + 525, + 568 + ], + "spans": [ + { + "bbox": [ + 302, + 392, + 525, + 568 + ], + "type": "text", + "content": "Recently, there has been increased interest in using language models for theorem proving in formal interactive theorem provers (ITP) (e.g., Polu and Sutskever (2020); Han et al. (2022); Polu et al. (2023); Jiang et al. (2022b,a); Lample et al. (2022)). Example ITPs include Lean (Moura et al., 2015), Isabelle (Paulson, 1994), Coq (Barras et al., 1999), and Metamath (Megill and Wheeler, 2019). To prove a theorem in an ITP, the theorem is stated in the ITP's programming language, then simplified by generating \"proof steps\" until it is reduced to known facts. The result is a sequence of steps that constitutes a verified proof." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 571, + 525, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 571, + 525, + 772 + ], + "spans": [ + { + "bbox": [ + 302, + 571, + 525, + 772 + ], + "type": "text", + "content": "Data sources for neural theorem proving in ITPs include interactive learning environments that interface with ITPs, and datasets derived from proofs in ITP libraries. For example, CoqGym (Yang and Deng, 2019) provides an interactive environment and 71K human-written proofs for the Coq ITP. For Isabelle, PISA (Jiang et al., 2021) enables interaction and provides a dataset of 183k proofs mined from the Isabelle standard library and Archive of Formal Proofs. For Lean, LeanStep (Han et al., 2022) provides a dataset of proof-steps from Lean's mathematical library along with auxiliary tasks, while Lean-Gym (Polu et al., 2023) provides an interactive REPL. The miniF2F (Zheng et al., 2022) benchmark aims to provide a shared benchmark" + } + ] + } + ], + "index": 11 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "text", + "content": "14625" + } + ] + } + ], + "index": 12 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 20 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 289, + 98 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 289, + 98 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 289, + 98 + ], + "type": "text", + "content": "across ITPs, consisting of 488 problem statements sourced from mathematical competitions." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 100, + 291, + 222 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 100, + 291, + 222 + ], + "spans": [ + { + "bbox": [ + 67, + 100, + 291, + 222 + ], + "type": "text", + "content": "Other resources provide proxy environments or tasks. For example, INT (Wu et al., 2021c) provide a synthetic proving environment to measure six different types of generalization. Li et al. construct IsarStep using the Isabelle Archive of Formal Proofs, and propose a task of filling in a missing intermediate proposition. Early applications of deep learning for formal theorem proving focus on selecting relevant premises (Alemi et al., 2016)." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "spans": [ + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "type": "text", + "content": "Informal theorem proving presents an alternative medium for theorem proving, in which statements and proofs are written in the mixture of natural language and symbols used in \"standard\" mathematics (e.g., in " + }, + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "type": "inline_equation", + "content": "\\mathrm{LATEX}" + }, + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "type": "text", + "content": "), and are checked for correctness by humans. Early work focuses on selecting relevant premises (Ferreira and Freitas, 2020b,a). Welleck et al. (2021) develop NaturalProofs, a large-scale dataset of " + }, + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "type": "inline_equation", + "content": "32\\mathrm{k}" + }, + { + "bbox": [ + 69, + 223, + 291, + 412 + ], + "type": "text", + "content": " informal mathematical theorems, definitions, and proofs, and provide a benchmark for premise selection via retrieval and generation tasks. Welleck et al. (2022a) adapt NaturalProofs for full proof generation, and provide a human evaluation protocol and proxy automatic metrics." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 414, + 291, + 549 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 414, + 291, + 549 + ], + "spans": [ + { + "bbox": [ + 67, + 414, + 291, + 549 + ], + "type": "text", + "content": "An emerging area of research aims to combine elements of informal and formal theorem proving. For example, Wu et al. (2022b) explore translating informal statements into formal statements, while Jiang et al. (2022a) release a new version of the miniF2F benchmark augmented with informal statements and proofs, which we refer to as miniF2F+informal. Jiang et al. (2022a) explore translating provided (or generated) informal proofs into formal proofs." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 68, + 564, + 226, + 577 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 564, + 226, + 577 + ], + "spans": [ + { + "bbox": [ + 68, + 564, + 226, + 577 + ], + "type": "text", + "content": "A.3 Geometry Problem Solving" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 584, + 291, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 584, + 291, + 772 + ], + "spans": [ + { + "bbox": [ + 67, + 584, + 291, + 772 + ], + "type": "text", + "content": "Automated geometry problem solving (GPS) is also a long-standing AI task in mathematical reasoning research (Gelernter et al., 1960; Wen-Tsun, 1986; Chou et al., 1996; Ye et al., 2008) and has attracted much attention in recent years. Different from a math word problem, a geometry problem consists of a textual description in natural language and a geometric diagram. As shown in Figure 2, the multimodal inputs describe the entities, attributes, and relationships of geometric elements, and the goal is to find the numeric solution to an unknown variable. GPS is a challenging task for deep learning methods due to the complex skills it requires. It involves the ability to parse multimodal informa" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 302, + 71, + 524, + 98 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 524, + 98 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 524, + 98 + ], + "type": "text", + "content": "tion, perform symbolic abstraction, utilize theorem knowledge, and conduct quantitative reasoning." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 302, + 100, + 526, + 302 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 100, + 526, + 302 + ], + "spans": [ + { + "bbox": [ + 302, + 100, + 526, + 302 + ], + "type": "text", + "content": "Some early datasets are proposed to facilitate research in this domain (Seo et al., 2015; Alvin et al., 2017; Sachan et al., 2017; Sachan and Xing, 2017). However, these datasets are relatively small or not publicly available, which limits the development of deep learning methods. In response to this limitation, Lu et al. create the Geometry3K dataset, which consists of 3,002 multi-choice geometry problems with unified logic form annotations for the multimodal inputs. More recently, larger-scale datasets such as GeoQA (Chen et al., 2021a), GeoQA+ (Cao and Xiao, 2022), and UniGeo (Chen et al., 2022a) have been introduced and are annotated with programs that can be learned by neural solvers and executed to obtain the final answers." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 317, + 456, + 330 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 317, + 456, + 330 + ], + "spans": [ + { + "bbox": [ + 302, + 317, + 456, + 330 + ], + "type": "text", + "content": "A.4 Math Question Answering" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 337, + 525, + 540 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 337, + 525, + 540 + ], + "spans": [ + { + "bbox": [ + 302, + 337, + 525, + 540 + ], + "type": "text", + "content": "Numerical reasoning is a core ability within human intelligence and plays an important role in many NLP tasks. Aside from theorem proving and grade-level math word problem solving, there is a wide range of question answering (QA) benchmarks that center around mathematical reasoning. In this work, we refer to these tasks as math question answering (MathQA). A large number of datasets have been presented recently. For example, QuaRel (Tafjord et al., 2019) is a dataset of diverse story questions that involve 19 different types of quantities. McTaco (Zhou et al., 2019) studies temporal commonsense problems, while Fermi (Kalyan et al., 2021) studies Fermi problems whose answers can only be approximately estimated." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 542, + 525, + 757 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 542, + 525, + 757 + ], + "spans": [ + { + "bbox": [ + 302, + 542, + 525, + 757 + ], + "type": "text", + "content": "Recent studies have shown that state-of-the-art mathematical reasoning systems might suffer from brittleness in reasoning, in that the models rely on spurious signals and plug-and-chug calculations in the specific dataset to achieve \"satisfactory\" performance (Hendrycks et al., 2021b; Mishra et al., 2022b). To address this issue, new benchmarks are proposed from various aspects. The Mathematics dataset (Saxton et al., 2020) consists of many different types of mathematics problems, covering arithmetic, algebra, probability, and calculus. The dataset allows for measuring the algebraic generalization ability of a model. Similarly, MATH (Hendrycks et al., 2021b) consists of challenging competition mathematics to measure the problem-solving ability of models in complex scenarios." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 314, + 760, + 524, + 772 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 760, + 524, + 772 + ], + "spans": [ + { + "bbox": [ + 314, + 760, + 524, + 772 + ], + "type": "text", + "content": "Some work incorporates tabular contexts in the" + } + ] + } + ], + "index": 11 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 312, + 791 + ], + "type": "text", + "content": "14626" + } + ] + } + ], + "index": 12 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 21 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 293, + 275 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 293, + 275 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 293, + 275 + ], + "type": "text", + "content": "question inputs. For example, FinQA (Chen et al., 2021c), TAT-QA (Zhu et al., 2021), and MultiHiertt (Zhao et al., 2022) collect questions that require both table understanding and numeric reasoning to answer. Others, instead, present large-scale unified benchmarks for mathematical reasoning (Mishra et al., 2022b,a; Chen et al., 2023). NumGLUE (Mishra et al., 2022b) is a multi-task benchmark with the goal of evaluating the performance of models on eight different tasks. Mishra et al. 2022a push this direction further and presents Lila, which consists of 23 mathematical reasoning tasks, spanning a wide range of mathematics topics, linguistic complexity, question formats, and background knowledge requirements." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 283, + 235, + 296 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 283, + 235, + 296 + ], + "spans": [ + { + "bbox": [ + 67, + 283, + 235, + 296 + ], + "type": "text", + "content": "A.5 Other Quantitative Problems" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 301, + 291, + 421 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 301, + 291, + 421 + ], + "spans": [ + { + "bbox": [ + 67, + 301, + 291, + 421 + ], + "type": "text", + "content": "Numbers are an integral part of our daily lives, and we humans reason with numbers in a variety of tasks, such as understanding news, reports, elections, and markets. This has led many in the community to question whether AI systems can effectively perform quantitative reasoning in everyday scenarios. To this end, various benchmarks have been developed to evaluate the capabilities of AI systems in this area." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 423, + 292, + 748 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 423, + 292, + 748 + ], + "spans": [ + { + "bbox": [ + 67, + 423, + 292, + 748 + ], + "type": "text", + "content": "Diagrams, such as figures, charts, and plots, are essential media that convey large amounts of information in a concise way. FigureQA (Kahou et al., 2018), DVQA (Kafle et al., 2018), MNS (Zhang et al., 2020c), PGDP5K (Hao et al., 2022), and GeoRE (Yu et al., 2021a), are released to investigate models' abilities to reason about quantitative relationships among entities grounded in diagrams. NumerSense (Lin et al., 2020), instead, examines whether and to what extent existing pre-trained language models can induce numerical commonsense knowledge. EQUATE (Ravichander et al., 2019) formalizes aspects of quantitative reasoning in a natural language inference framework. Quantitative reasoning can appear frequently in specific domains like finance, science, and programming. For instance, the ConvFinQA (Chen et al., 2022c) targets numerical reasoning over financial reports in a conversational question answering format. ScienceQA (Lu et al., 2022a) involves numerical reasoning in scientific domains, while P3 (Schuster et al., 2021) studies the function inference ability of deep learning models to find a valid input which makes the given program return True." + } + ] + } + ], + "index": 3 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 313, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 313, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 313, + 791 + ], + "type": "text", + "content": "14627" + } + ] + } + ], + "index": 4 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 22 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 68, + 68, + 533, + 723 + ], + "blocks": [ + { + "bbox": [ + 68, + 68, + 533, + 723 + ], + "lines": [ + { + "bbox": [ + 68, + 68, + 533, + 723 + ], + "spans": [ + { + "bbox": [ + 68, + 68, + 533, + 723 + ], + "type": "table", + "html": "
DatasetTaskSizeInputOutputRationaleDomain
Verb395 (2014)MWP395QuestionNumberEquationMath
Alg514 (2014)MWP514QuestionNumberEquationMath
IL (2015)MWP-QuestionNumberEquationMath
SingleEQ (2015)MWP508QuestionNumberEquationMath
DRAW (2015)MWP1,000QuestionNumberEquationMath
Dolphin1878 (2015)MWP1,878QuestionNumberEquationMath
Dolphin18K (2016)MWP18,460QuestionNumberEquationMath
MAWPS (2016)MWP3,320QuestionNumberEquationMath
AllArith (2017)MWP831QuestionNumberEquationMath
DRAW-1K (2017)MWP1,000QuestionNumberEquationMath
Math23K (2017)MWP23,162QuestionNumberEquationMath
AQuA (2017)MWP100,000QuestionOptionNatural languageMath
Aggregate (2018)MWP1,492QuestionNumberEquationMath
MathQA (2019)MWP37,297QuestionNumberProgramMath
ASDiv (2020)MWP2,305QuestionNumberEquationMath
HMWP (2020)MWP5,470QuestionNumberEquationMath
Ape210K (2020)MWP210,488QuestionNumberEquationMath
SVAMP (2021)MWP1,000QuestionNumberEquationMath
GSM8K (2021)MWP8,792QuestionNumberNatural languageMath
IconQA (2021b)MWP107,439Figure+QuestionOption+Text spanXMath
MathQA-Python (2021)MWP23,914QuestionNumberPython programMath
ArMATH (2022)MWP6,000QuestionNumberEquationMath
TabMWP (2022b)MWP38,431Table+QuestionOption+NumberNatural languageMath
MML (2015)TP57,882StatementProof stepsXMath
HolStep (2017)TP2,209,076StatementProof stepsXMath
Gamepad (2019)TP-StatementProof stepsXMath
CoqGym (2019)TP71,000StatementProof stepsXMath
HOList (2019)TP29,462StatementProof stepsXMath
IsarStep (2021)TP860,000StatementProof stepsXMath
PISA (2021)TP183,000StatementProof stepsXMath
INT (2021c)TP-StatementProof stepsXMath
NaturalProofs (2021)TP32,000StatementProof stepsXMath
NaturalProofs-Gen (2022a)TP14,500StatementProof stepsXMath
miniF2F (2022)TP488StatementProof stepsXMath
miniF2F+informal (2022a)TP488StatementProof stepsXMath
LeanStep (2022)TP21,606,000StatementProof stepsXMath
GEOS (2015)GPS186Figure+QuestionOptionXGeometry
GeoShader (2017)GPS102Figure+QuestionNumberXGeometry
GEOS++ (2017)GPS1,406Figure+QuestionNumberXGeometry
GEOS-OS (2017)GPS2,235Figure+QuestionOptionDemonstrationGeometry
Geometry3K (2021a)GPS3,002Figure+QuestionOptionLogical formGeometry
GeoQA (2021a)GPS4,998Figure+QuestionOptionProgramGeometry
GeoQA+ (2022)GPS12,054Figure+QuestionOptionProgramGeometry
UniGeo (2022a)GPS/TP14,541Figure+QuestionOptionProgramGeometry
Quarel (2019)MathQA2,771QuestionOptionLogical formMath
McTaco (2019)MathQA13,225Text+QuestionOptionXTime
DROP (2019)MathQA96,567Passage+QuestionNumber+Text spanXMath
Mathematics (2020)MathQA2,010,000QuestionFree-formNumberMath
FinQA (2021c)MathQA8,281Text+Table+QNumberProgramFinance
Fermi (2021)MathQA11,000QuestionNumberProgram+FactMath
MATH (2021b)MathQA12,500QuestionNumberNatural languageMath
TAT-QA (2021)MathQA16,552Text+Table+QNumber+Text spanXFinance
AMPS (2021b)MathQA5,000,000Question-LATEXMath
MultiHiertt (2022)MathQA10,440Text+Table+QNumber+Text spanExpressionFinance
NumGLUE (2022b)MathQA101,835Text+QuestionNumber+Text spanXMath
Lila (2022a)MathQA134,000Text+QuestionFree-formPython programMath
FigureQA (2018)VQA1,000,000+Figure+QuestionBinaryXMath
DVQA (2018)VQA3,487,194Figure+QuestionText spanNumber+Text spanMath
DREAM (2019)ConvQA10,197Dialog+QuestionOptionXMath
EQUATE (2019)NLI-Premise+HypothesisBinaryXMath
NumerSense (2020)Filling13,600Masked questionWordXMath
MNS (2020c)IQ Test-FigureNumberXMath
P3 (2021)Puzzle397TextProgramXMath
NOAHQA (2021)ConvQA21,347Dialog+QuestionText spanReasoning graphMath
ConvFinQA (2022c)ConvQA3,892Report+Dialog+QNumberExpressionMath
PGDP5K (2022)Parsing5,000Figure+QuestionNumberXGeometry
GeoRE (2022a)Parsing12,901Figure+QuestionNumberXGeometry
ScienceQA (2022a)VQA21,208Context+QuestionOptionNatural languageScience
", + "image_path": "963c831806d2554a2c4aff2b7370a6050bd857131f2df6c534a972d46dfea850.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "bbox": [ + 170, + 730, + 422, + 741 + ], + "lines": [ + { + "bbox": [ + 170, + 730, + 422, + 741 + ], + "spans": [ + { + "bbox": [ + 170, + 730, + 422, + 741 + ], + "type": "text", + "content": "Table 7: A summarization of mathematical reasoning datasets." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14628" + } + ] + } + ], + "index": 2 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 23 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 68, + 207, + 526, + 597 + ], + "blocks": [ + { + "bbox": [ + 68, + 207, + 526, + 597 + ], + "lines": [ + { + "bbox": [ + 68, + 207, + 526, + 597 + ], + "spans": [ + { + "bbox": [ + 68, + 207, + 526, + 597 + ], + "type": "table", + "html": "
PaperTaskProblemNetworkEncodeDecodATTDescription
DNS (Wang et al., 2017)MWPGenerationSeq2SeqGRULSTMXThe first deep MWP solver
AnsRat (Ling et al., 2017)MWPGenerationSeq2SeqLSTMLSTMXTrained with staged back-propagation
Math-EN (Wang et al., 2018a)MWPGenerationSeq2SeqBiLSTMLSTMA standard Seq2Seq model with attention
CASS (Huang et al., 2018)MWPGenerationSeq2SeqBiGRUBiGRUCopy and alignment with RL
S-Aligned (Chiang and Chen, 2019)MWPGenerationSeq2SeqBiLSTMLSTMOperating symbols
T-RNN (Wang et al., 2019)MWPGenerationSeq2SeqBiLSTMBiLSTMPredicting a tree-structure math template
GROUP-ATT (Li et al., 2019)MWPGenerationSeq2SeqBiLSTMLSTMGroup attention
SMART (Hong et al., 2021b)MWPGenerationSeq2Seq--XExplicitly incorporating values
SelfAtt (Robaidek et al., 2018)GPSClassificationSeq2SeqBiLSTM-Multi-hop self-attention
QuaSP+ (Tajord et al., 2019)MathQAGenerationSeq2SeqBiLSTMLSTMXAdopting attributed grammar
AST-Dec (Liu et al., 2019a)MWPGenerationSeq2TreeBiLSTMTreeUsing prefix order decoding
GTS (Xie and Sun, 2019)MWPGenerationSeq2TreeBiGRUTreeA goal-driven tree-structured approach
KA-S2T (Wu et al., 2020)MWPGenerationSeq2TreeBiLSTMTreeA knowledge-aware method
TSN-MD (Zhang et al., 2020a)MWPGenerationSeq2TreeBiGRUTreeA teacher-student network
T-LSTM (Zaporojets et al., 2021)MWPGenerationSeq2TreeBiLSTMTreeXA child-sum tree-LSTM model
NT-LSTM (Zaporojets et al., 2021)MWPGenerationSeq2TreeBiLSTMTreeXAn N-ary tree-LSTM model
NS-Solver (Qin et al., 2021)MWPGenerationSeq2TreeBiGRUTreeA neural-symbolic solver with programs
NumS2T (Wu et al., 2021b)MWPGenerationSeq2TreeBiLSTMTreeExplicitly incorporating values
HMS (Lin et al., 2021)MWPGenerationSeq2TreeGRUTreeA word-clause-problem encoder
LBF (Hong et al., 2021a)MWPGenerationSeq2TreeBiGRUTreeA learning-by-fixing (LBF) framework
Seq2DAG (Cao et al., 2021)MWPGenerationSeq2GraphGRUGraphXA direct acyclic graph (DAG) structure
Graph2Tree (Zhang et al., 2020b)MWPGenerationGraph2TreeGraphTreeXGenerating better solution expressions
Multi-E/D (Shen and Jin, 2020)MWPGenerationGraph2TreeGraphTreeA graph encoder and a tree-bad decoder
Graph2Tree (Li et al., 2020b)MWPGenerationGraph2TreeGraphTreeA graph-to-tree neural network
EEH-G2T (Wu et al., 2021a)MWPGenerationGraph2TreeGraphTreeXA hierarchical graph-to-tree model
ASTactic (Yang and Deng, 2019)TPGenerationTree2SeqTreeLSTMGRUGenerating tactics as programs
MathDQN (Wang et al., 2018b)MWPSearchDQN--XRL with a deep Q-network (DQN)
DDT (Meng and Rumshisky, 2019)MWPGenerationTransformerTrmTrmA Transformer-based model
DeepMath (Alemi et al., 2016)TPClassificationCNNCNN-XThe first deep large scale theorem prover
Holophrasm (Whalen, 2016)TPClassificationBiGRUBiGRU-XA neural prover for higher-order logic
CNNTP (Loos et al., 2017)TPClassificationCNNCNN-XA CNN-based theorem prover
WaveNetTP (Loos et al., 2017)TPClassificationWaveNetWaveNet-XA WaveNet-based theorem prover
DeepHOL (Bansal et al., 2019)TPGenerationWaveNetWaveNet-XA neural theorem prover with RL
NGS (Chen et al., 2021a)GPSGenerationVQALSTM*LSTMThe first deep geometry solver
PGDPNet (Zhang et al., 2022)ParsingGenerationGNN--XA neural diagram parser with GNN
", + "image_path": "642e6497d07fe9f4f9f4915424f2587bcc4435ccfd903968a8e2d2195cb64cf1.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 607, + 525, + 631 + ], + "lines": [ + { + "bbox": [ + 67, + 607, + 525, + 631 + ], + "spans": [ + { + "bbox": [ + 67, + 607, + 525, + 631 + ], + "type": "text", + "content": "Table 8: A summarization of deep neural network models for mathematical reasoning. Encod: encoder, Decod: decoder, ATT: Attention. LSTM*: ResNet + LSTM, Trm: Transformer" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14629" + } + ] + } + ], + "index": 2 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 24 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 90, + 192, + 103 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 90, + 192, + 103 + ], + "spans": [ + { + "bbox": [ + 69, + 90, + 192, + 103 + ], + "type": "text", + "content": "A For every submission:" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 77, + 106, + 317, + 121 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 106, + 317, + 121 + ], + "spans": [ + { + "bbox": [ + 77, + 106, + 317, + 121 + ], + "type": "text", + "content": "A1. Did you describe the limitations of your work?" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 89, + 121, + 230, + 134 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 121, + 230, + 134 + ], + "spans": [ + { + "bbox": [ + 89, + 121, + 230, + 134 + ], + "type": "text", + "content": "Limitations Section on page 10." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 77, + 142, + 329, + 157 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 142, + 329, + 157 + ], + "spans": [ + { + "bbox": [ + 77, + 142, + 329, + 157 + ], + "type": "text", + "content": "A2. Did you discuss any potential risks of your work?" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 89, + 158, + 230, + 170 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 158, + 230, + 170 + ], + "spans": [ + { + "bbox": [ + 89, + 158, + 230, + 170 + ], + "type": "text", + "content": "Limitations Section on page 10." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 77, + 179, + 414, + 193 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 179, + 414, + 193 + ], + "spans": [ + { + "bbox": [ + 77, + 179, + 414, + 193 + ], + "type": "text", + "content": "A3. Do the abstract and introduction summarize the paper's main claims?" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 91, + 194, + 135, + 205 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 91, + 194, + 135, + 205 + ], + "spans": [ + { + "bbox": [ + 91, + 194, + 135, + 205 + ], + "type": "text", + "content": "Section 1." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 77, + 216, + 398, + 229 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 216, + 398, + 229 + ], + "spans": [ + { + "bbox": [ + 77, + 216, + 398, + 229 + ], + "type": "text", + "content": "A4. Have you used AI writing assistants when working on this paper?" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 91, + 230, + 138, + 242 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 91, + 230, + 138, + 242 + ], + "spans": [ + { + "bbox": [ + 91, + 230, + 138, + 242 + ], + "type": "text", + "content": "Left blank." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 253, + 290, + 266 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 253, + 290, + 266 + ], + "spans": [ + { + "bbox": [ + 69, + 253, + 290, + 266 + ], + "type": "text", + "content": "B Did you use or create scientific artifacts?" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 79, + 270, + 127, + 283 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 79, + 270, + 127, + 283 + ], + "spans": [ + { + "bbox": [ + 79, + 270, + 127, + 283 + ], + "type": "text", + "content": "Left blank." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 77, + 292, + 315, + 306 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 292, + 315, + 306 + ], + "spans": [ + { + "bbox": [ + 77, + 292, + 315, + 306 + ], + "type": "text", + "content": "B1. Did you cite the creators of artifacts you used?" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 89, + 306, + 208, + 319 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 306, + 208, + 319 + ], + "spans": [ + { + "bbox": [ + 89, + 306, + 208, + 319 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 77, + 328, + 463, + 342 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 328, + 463, + 342 + ], + "spans": [ + { + "bbox": [ + 77, + 328, + 463, + 342 + ], + "type": "text", + "content": "B2. Did you discuss the license or terms for use and / or distribution of any artifacts?" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 90, + 343, + 208, + 355 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 343, + 208, + 355 + ], + "spans": [ + { + "bbox": [ + 90, + 343, + 208, + 355 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 77, + 364, + 524, + 417 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 364, + 524, + 417 + ], + "spans": [ + { + "bbox": [ + 77, + 364, + 524, + 417 + ], + "type": "text", + "content": "B3. Did you discuss if your use of existing artifact(s) was consistent with their intended use, provided that it was specified? For the artifacts you create, do you specify intended use and whether that is compatible with the original access conditions (in particular, derivatives of data accessed for research purposes should not be used outside of research contexts)?" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 90, + 419, + 208, + 432 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 419, + 208, + 432 + ], + "spans": [ + { + "bbox": [ + 90, + 419, + 208, + 432 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 77, + 441, + 524, + 481 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 441, + 524, + 481 + ], + "spans": [ + { + "bbox": [ + 77, + 441, + 524, + 481 + ], + "type": "text", + "content": "B4. Did you discuss the steps taken to check whether the data that was collected / used contains any information that names or uniquely identifies individual people or offensive content, and the steps taken to protect / anonymize it?" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 90, + 482, + 208, + 495 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 482, + 208, + 495 + ], + "spans": [ + { + "bbox": [ + 90, + 482, + 208, + 495 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 77, + 504, + 524, + 531 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 504, + 524, + 531 + ], + "spans": [ + { + "bbox": [ + 77, + 504, + 524, + 531 + ], + "type": "text", + "content": "B5. Did you provide documentation of the artifacts, e.g., coverage of domains, languages, and linguistic phenomena, demographic groups represented, etc.?" + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 90, + 532, + 208, + 544 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 532, + 208, + 544 + ], + "spans": [ + { + "bbox": [ + 90, + 532, + 208, + 544 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 77, + 554, + 524, + 622 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 554, + 524, + 622 + ], + "spans": [ + { + "bbox": [ + 77, + 554, + 524, + 622 + ], + "type": "text", + "content": "B6. Did you report relevant statistics like the number of examples, details of train / test / dev splits, etc. for the data that you used / created? Even for commonly-used benchmark datasets, include the number of examples in train / validation / test splits, as these provide necessary context for a reader to understand experimental results. For example, small differences in accuracy on large test sets may be significant, while on small test sets they may not be." + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 90, + 623, + 208, + 634 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 623, + 208, + 634 + ], + "spans": [ + { + "bbox": [ + 90, + 623, + 208, + 634 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 69, + 644, + 293, + 657 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 644, + 293, + 657 + ], + "spans": [ + { + "bbox": [ + 69, + 644, + 293, + 657 + ], + "type": "text", + "content": "C Did you run computational experiments?" + } + ] + } + ], + "index": 24 + }, + { + "bbox": [ + 79, + 661, + 127, + 674 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 79, + 661, + 127, + 674 + ], + "spans": [ + { + "bbox": [ + 79, + 661, + 127, + 674 + ], + "type": "text", + "content": "Left blank." + } + ] + } + ], + "index": 25 + }, + { + "bbox": [ + 77, + 684, + 524, + 711 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 77, + 684, + 524, + 711 + ], + "spans": [ + { + "bbox": [ + 77, + 684, + 524, + 711 + ], + "type": "text", + "content": "C1. Did you report the number of parameters in the models used, the total computational budget (e.g., GPU hours), and computing infrastructure used?" + } + ] + } + ], + "index": 26 + }, + { + "bbox": [ + 90, + 712, + 208, + 724 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 712, + 208, + 724 + ], + "spans": [ + { + "bbox": [ + 90, + 712, + 208, + 724 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 27 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 79, + 71, + 258, + 84 + ], + "type": "header", + "angle": 0, + "lines": [ + { + "bbox": [ + 79, + 71, + 258, + 84 + ], + "spans": [ + { + "bbox": [ + 79, + 71, + 258, + 84 + ], + "type": "text", + "content": "ACL 2023 Responsible NLP Checklist" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 730, + 522, + 751 + ], + "type": "footer", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 730, + 522, + 751 + ], + "spans": [ + { + "bbox": [ + 67, + 730, + 522, + 751 + ], + "type": "text", + "content": "The Responsible NLP Checklist used at ACL 2023 is adopted from NAACL 2022, with the addition of a question on AI writing assistance." + } + ] + } + ], + "index": 28 + }, + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 781, + 312, + 791 + ], + "type": "text", + "content": "14630" + } + ] + } + ], + "index": 29 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 25 + }, + { + "para_blocks": [ + { + "bbox": [ + 76, + 71, + 523, + 97 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 71, + 523, + 97 + ], + "spans": [ + { + "bbox": [ + 76, + 71, + 523, + 97 + ], + "type": "text", + "content": "C2. Did you discuss the experimental setup, including hyperparameter search and best-found hyperparameter values?" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 89, + 99, + 208, + 111 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 99, + 208, + 111 + ], + "spans": [ + { + "bbox": [ + 89, + 99, + 208, + 111 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 76, + 121, + 525, + 161 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 121, + 525, + 161 + ], + "spans": [ + { + "bbox": [ + 76, + 121, + 525, + 161 + ], + "type": "text", + "content": "C3. Did you report descriptive statistics about your results (e.g., error bars around results, summary statistics from sets of experiments), and is it transparent whether you are reporting the max, mean, etc. or just a single run?" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 89, + 162, + 208, + 174 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 162, + 208, + 174 + ], + "spans": [ + { + "bbox": [ + 89, + 162, + 208, + 174 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 76, + 184, + 525, + 223 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 184, + 525, + 223 + ], + "spans": [ + { + "bbox": [ + 76, + 184, + 525, + 223 + ], + "type": "text", + "content": "C4. If you used existing packages (e.g., for preprocessing, for normalization, or for evaluation), did you report the implementation, model, and parameter settings used (e.g., NLTK, Spacy, ROUGE, etc.)?" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 89, + 225, + 208, + 238 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 225, + 208, + 238 + ], + "spans": [ + { + "bbox": [ + 89, + 225, + 208, + 238 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 247, + 522, + 261 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 247, + 522, + 261 + ], + "spans": [ + { + "bbox": [ + 67, + 247, + 522, + 261 + ], + "type": "text", + "content": "D Did you use human annotators (e.g., crowdworkers) or research with human participants?" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 79, + 264, + 127, + 277 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 79, + 264, + 127, + 277 + ], + "spans": [ + { + "bbox": [ + 79, + 264, + 127, + 277 + ], + "type": "text", + "content": "Left blank." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 76, + 286, + 525, + 313 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 286, + 525, + 313 + ], + "spans": [ + { + "bbox": [ + 76, + 286, + 525, + 313 + ], + "type": "text", + "content": "D1. Did you report the full text of instructions given to participants, including e.g., screenshots, disclaimers of any risks to participants or annotators, etc.?" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 89, + 315, + 208, + 327 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 315, + 208, + 327 + ], + "spans": [ + { + "bbox": [ + 89, + 315, + 208, + 327 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 76, + 336, + 525, + 376 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 336, + 525, + 376 + ], + "spans": [ + { + "bbox": [ + 76, + 336, + 525, + 376 + ], + "type": "text", + "content": "D2. Did you report information about how you recruited (e.g., crowdsourcing platform, students) and paid participants, and discuss if such payment is adequate given the participants' demographic (e.g., country of residence)?" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 89, + 378, + 208, + 391 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 378, + 208, + 391 + ], + "spans": [ + { + "bbox": [ + 89, + 378, + 208, + 391 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 76, + 400, + 525, + 439 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 400, + 525, + 439 + ], + "spans": [ + { + "bbox": [ + 76, + 400, + 525, + 439 + ], + "type": "text", + "content": "D3. Did you discuss whether and how consent was obtained from people whose data you're using/curating? For example, if you collected data via crowdsourcing, did your instructions to crowdworkers explain how the data would be used?" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 89, + 441, + 208, + 454 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 441, + 208, + 454 + ], + "spans": [ + { + "bbox": [ + 89, + 441, + 208, + 454 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 76, + 462, + 520, + 476 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 462, + 520, + 476 + ], + "spans": [ + { + "bbox": [ + 76, + 462, + 520, + 476 + ], + "type": "text", + "content": "D4. Was the data collection protocol approved (or determined exempt) by an ethics review board?" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 89, + 476, + 208, + 489 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 476, + 208, + 489 + ], + "spans": [ + { + "bbox": [ + 89, + 476, + 208, + 489 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 76, + 498, + 525, + 524 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 76, + 498, + 525, + 524 + ], + "spans": [ + { + "bbox": [ + 76, + 498, + 525, + 524 + ], + "type": "text", + "content": "D5. Did you report the basic demographic and geographic characteristics of the annotator population that is the source of the data?" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 89, + 527, + 208, + 539 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 527, + 208, + 539 + ], + "spans": [ + { + "bbox": [ + 89, + 527, + 208, + 539 + ], + "type": "text", + "content": "Not applicable. Left blank." + } + ] + } + ], + "index": 17 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 284, + 780, + 311, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 284, + 780, + 311, + 791 + ], + "spans": [ + { + "bbox": [ + 284, + 780, + 311, + 791 + ], + "type": "text", + "content": "14631" + } + ] + } + ], + "index": 18 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 26 + } + ], + "_backend": "vlm", + "_version_name": "2.6.4" +} \ No newline at end of file